648 followers 206 articoli a settimana
[R] I made an app to predict ICML paper acceptance from reviews

https://www.norange.io/projects/paper_scorer/ A couple of years ago, u/programmerChilli analyzed ICLR 2019 reviews data and trained a model that rather accurately predicted acceptance results for NeurIPS. I've decided to continue this analysis and trained a model (total ~6000 parameters) on newer NeurIPS reviews, which has twice as many reviews compared...

Wed Apr 24, 2024 17:24
[R] SpaceByte: Towards Deleting Tokenization from Large Language Modeling - Rice University 2024 - Practically the same performance as subword tokenizers without their many downsides!

Paper: https://arxiv.org/abs/2404.14408 Github: https://github.com/kjslag/spacebyte Abstract: Tokenization is widely used in large language models because it significantly improves performance. However, tokenization imposes several disadvantages, such as performance biases, increased adversarial vulnerability, decreased character-level modeling performance,...

Wed Apr 24, 2024 17:24
[D] how can a ChatGPT generate such a response?

I just typed 'a' to the prompt, and it answered like this https://preview.redd.it/edof9ug1sewc1.png?width=1136&format=png&auto=webp&s=0a122354b1bbd297187174256837122052c68fd2 I didn't provide any context (or prompt) except a single letter 'a'. However, it generated a reasonable answer. Up to my knowledge, LLM is trained by the following...

Wed Apr 24, 2024 14:25
[D] Keeping track of models and their associated metadata.

I am starting to accumulate a large number of models for a project I am working on, many of these models are old which I am keeping for archival sake, and many are fine tuned from other models. I am wondering if there is an industry standard way of dealing with this, in particular I am looking for the following: Information about parameters used to...

Wed Apr 24, 2024 14:25
[D] Deploy the fine-tuned Mistral 7B model using the Hugging Face library

I followed the tutorial provided at https://www.datacamp.com/tutorial/mistral-7b-tutorial and now seek methods to deploy the model for faster inference using Hugging Face and Gradio. Could anyone please share a guide notebook or article for reference? Any help would be appreciated. submitted by /u/Future-Outcome3167 [link] [comments]

Wed Apr 24, 2024 14:25
[N] Perplexity is valued at $1 Billion in Funding Round!

AI News Today – April 23rd, 2024 Here’s a quick look at what’s happening in the world of AI today: Perplexity, an AI Search Startup, Valued at $1 Billion in Funding Round Microsoft Introduces Smaller AI Model to Attract Wider Client Base Google-Backed Glance Pilots Android Lockscreen Platform in US AI Predicts Political Orientations from Blank Faces,...

Wed Apr 24, 2024 11:25

Costruisci il tuo feed di notizie

Pronto a fare un tentativo?
Inizia una prova di 14 giorni, nessuna carta di credito richiesta.

Crea account