style
- Asia > China > Shaanxi Province > Xi'an (0.04)
- North America > United States (0.04)
- Asia > Middle East > Israel (0.04)
Appendix AImplementationdetails
The encoder contains three linear layers with output size [d,k,k], each but the last layer is followed by batch normalization, witheps = 0.00005 and momentum=0.1,andtheReLU Thedecoder contains threelinear layers withoutput size [k,k,d] where each but the last layer contains a Batch normlization and the ReLu activation similar asabove. Following the standard linear evaluation procedure inself-supervised learning works (32;34),we used an one linear layer network as the linear decoder for the decoding accuracy. We used the neural activity dataset that is collected from two rhesus macaque monkeys (Chewie and Mihi). They were trained to move the computer cursor to reach a target on a screen.
Self-Supervised Learning with Data Augmentations Provably Isolates Content from Style
Self-supervised representation learning has shown remarkable success in a number of domains. A common practice is to perform data augmentation via hand-crafted transformations intended to leave the semantics of the data invariant. We seek to understand the empirical success of this approach from a theoretical perspective. We formulate the augmentation process as a latent variable model by postulating a partition of the latent representation into a content component, which is assumed invariant to augmentation, and a style component, which is allowed to change. Unlike prior work on disentanglement and independent component analysis, we allow for both nontrivial statistical and causal dependencies in the latent space. We study the identifiability of the latent representation based on pairs of views of the observations and prove sufficient conditions that allow us to identify the invariant content partition up to an invertible mapping in both generative and discriminative settings. We find numerical simulations with dependent latent variables are consistent with our theory. Lastly, we introduce Causal3DIdent, a dataset of high-dimensional, visually complex images with rich causal dependencies, which we use to study the effect of data augmentations performed in practice.
- Asia > China > Shaanxi Province > Xi'an (0.04)
- North America > United States (0.04)
- Asia > Middle East > Israel (0.04)
Can Large Language Models Understand Symbolic Graphics Programs?
Qiu, Zeju, Liu, Weiyang, Feng, Haiwen, Liu, Zhen, Xiao, Tim Z., Collins, Katherine M., Tenenbaum, Joshua B., Weller, Adrian, Black, Michael J., Schölkopf, Bernhard
Assessing the capabilities of large language models (LLMs) is often challenging, in part, because it is hard to find tasks to which they have not been exposed during training. We take one step to address this challenge by turning to a new task: focusing on symbolic graphics programs, which are a popular representation for graphics content that procedurally generates visual data. LLMs have shown exciting promise towards program synthesis, but do they understand symbolic graphics programs? Unlike conventional programs, symbolic graphics programs can be translated to graphics content. Here, we characterize an LLM's understanding of symbolic programs in terms of their ability to answer questions related to the graphics content. This task is challenging as the questions are difficult to answer from the symbolic programs alone -- yet, they would be easy to answer from the corresponding graphics content as we verify through a human experiment. To understand symbolic programs, LLMs may need to possess the ability to imagine how the corresponding graphics content would look without directly accessing the rendered visual content. We use this task to evaluate LLMs by creating a large benchmark for the semantic understanding of symbolic graphics programs. This benchmark is built via program-graphics correspondence, hence requiring minimal human efforts. We evaluate current LLMs on our benchmark to elucidate a preliminary assessment of their ability to reason about visual scenes from programs. We find that this task distinguishes existing LLMs and models considered good at reasoning perform better. Lastly, we introduce Symbolic Instruction Tuning (SIT) to improve this ability. Specifically, we query GPT4-o with questions and images generated by symbolic programs. Such data are then used to finetune an LLM. We also find that SIT data can improve the general instruction following ability of LLMs.
- Europe > Germany > Baden-Württemberg > Tübingen Region > Tübingen (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)
- Europe > Germany > Berlin (0.04)
- (2 more...)
- Education (1.00)
- Information Technology (0.67)
AI Chatbot Writes 'In the Style of Nick Cave,' and Nick Cave is Heated – Rolling Stone
Nick Cave, the Bad Seeds frontman whose songs are tinged with a healthy dose of death, forlorn love, and religion, is no fan of ChatGPT's lyrical ambitions. The popular AI bot has drawn both praise and concern for its ability to generate conversational and nuanced text responses in simple, clean sentences. Since its release in November by the artificial intelligence lab OpenAI, ChatGPT has written everything from sitcom scripts to literature essays to, now, rather convincing rock songs. This has left people worried about the ramifications for industries across the creative spectrum, and one of those people is Cave himself. In his latest The Red Hand Files newsletter, Cave took on the subject of AI generated music.
- Media > Music (0.75)
- Leisure & Entertainment (0.75)
John Madden returning to cover of Madden NFL 23 video game
Fox News Flash top headlines are here. Check out what's clicking on Foxnews.com. For the first time in two decades, late football legend John Madden will grace the cover of a Madden NFL video game. EA Sports on Wednesday announced that the Hall of Fame coach, who died in December, will appear on the cover of all three editions of this year's Madden NFL 23 video game. The covers will include him in different parts of his life, including as a coach and as a broadcaster.
- North America > United States > Minnesota (0.07)
- North America > United States > Michigan > Wayne County > Detroit (0.05)
- North America > United States > California > Los Angeles County > Pasadena (0.05)
- Leisure & Entertainment > Sports > Football (1.00)
- Leisure & Entertainment > Games > Computer Games (1.00)
iMovie 3.0 for the iPhone and iPad makes it easier to create your first video
While Apple's iMovie software has always been a fairly powerful piece of movie-editing software, its main appeal is that it's a free and simple way to start making your own videos. But iMovie for the iPhone and iPad offers surprisingly little in the way of direction when you open it -- it's up to the user to figure out how to structure their movie. Apple is changing that significantly today with iMovie 3.0 for iOS and iPadOS, a free download that's available now. The software features two new creation modes, Magic Movie and Storyboards, to help people build videos for the first time. Magic Movie automates a lot of the video creation process while still leaving room for plenty of customization, assuming you want to dig into the details.