I manipulated the hidden state values obtained from the llama-2 model after feeding it a certain input, let's call it Input_1. Now, I want to examine the output (causal output) it produces from this. My hypothesis is that it should correspond to a different input, let's call it Input_2, which would yield a distinct output from the initial input. I got...
Hi, I'm a lead data analyst trying to tweak my career direction. I've got some basic exposure to Web Development and Deep Learning. Which direction should I take, considering AI can disrupt future job market in either of them or both? submitted by /u/Sufficient-Result987 [link] [comments]
I am working on a research projects which involves experimenting with RAGs. I want to run the models first to get an understanding of how the whole pipeline works. I found some datasets on HuggingFace (such as https://huggingface.co/datasets/explodinggradients/WikiEval). My understanding of RAGs is that I should be given a datastore, and then I perform...
Did you guys ever feel that the same open source llm is giving slightly different answers on different playgrounds.. Like if you use llama 70 b on perplexity and groq notice the difference Can someone tell me why is it.. submitted by /u/IntentionNo5258 [link] [comments]
Hi everyone, is there any known set of hyperparameters for training a diffusion model on CIFAR10 or any other famous datasets primarily for reconstruction loss ? submitted by /u/sidney_lumet [link] [comments]
Title says it all. submitted by /u/darkknight-6 [link] [comments]
Build your own newsfeed
Ready to give it a go?
Start a 14-day trial, no credit card required.