Comments (19)
The sequences are all memorized in that, if you were to prompt them with a longer prefix than we're giving you, the suffix would be memorized. We've provided only a shorter prefix so we still believe it should be possible to reach relatively high numbers. We can't guarantee it's 100% solvable, but expect it should be much better than the baseline.
from lm-extraction-benchmark.
The logical reasoning for the query to start with "dataset is unknown" for the infinite possible outcomes. This allows for predictive queries to exist in coding. Your thoughts?
from lm-extraction-benchmark.
Pre- or suf- fix are traditional markers in written language script. They do. not belong in code, it disables the potentiality of data defining best outcomes. Your thoughts are welcome.
from lm-extraction-benchmark.
I'm sorry @heitikei I don't understand what you're trying to say and don't understand how it relates to either machine learning or language model extraction in particular. Can you rephrase your question?
from lm-extraction-benchmark.
@carlini sorry I am naturally cryptic, looking for the concept in english to explain what I see. Remove identification tags in queries will allow machines, AI and neural networks to perform as imagined......, ? ? ? ?
from lm-extraction-benchmark.
Reinforcement training is no longer useful
from lm-extraction-benchmark.
Hey @carlini, assuming that the dataset on the ETH Zürich GitHub repo is correct when giving in the preprefix + prefix to the language model, it only correctly guesses the suffix 56% of the time (on the last 1000 samples). Am I doing something wrong? Shouldn't there be a 100% accuracy if the sequence is memorized?
from lm-extraction-benchmark.
@carlini If I understand correctly if instead of feeding in 50 tokens as the prefix, we feed in 50+x tokens then the LM is guaranteed to spit out the next 50-x tokens?
from lm-extraction-benchmark.
Agreed. Match.
from lm-extraction-benchmark.
@akul-goyal Not exactly. There is a longer sequence (Call it S) that was partitioned to form (P1 + P2 + Q)
If you feed P1 + P2
then you get Q
. We give you P2
, but not P1
. So P2
is "almost" enough to give you Q
but not quite.
from lm-extraction-benchmark.
@heitikei Please open a new issue with an any questions you have. As far as I can tell they are not related to this issue here.
from lm-extraction-benchmark.
New issue. Close #8.
from lm-extraction-benchmark.
No; I will leave #8 open until we resolve the issue with @akul-goyal. You can open a new issue for your questions which appear different.
from lm-extraction-benchmark.
Thank you. Finding the concept. Sorry, biological unit is slow.
from lm-extraction-benchmark.
sorry, @carlini I just wanted to confirm from your comment you are giving us P2. Because you wrote "P1 is "almost" enough to give you Q but not quite" and I am not sure if that was an error or not
from lm-extraction-benchmark.
Also in the case we are being given P2, could you tell us how many tokens P1 is? Or does that vary?
from lm-extraction-benchmark.
Sorry yes! I meant P2 there. I've edited it for correctness for future readers.
P1 is length 50 tokens, P2 is length 50 tokens, and Q is length 50 tokens.
from lm-extraction-benchmark.
what is the performance indicator? assume guess "ing" out side of trained data set. This is good. Why is this good? Apologies, I am a natural language programmer and think -> speak. I struggle with emotional qualities of good communication. :/
from lm-extraction-benchmark.
Never mind, was not querying the model correctly!
from lm-extraction-benchmark.
Related Issues (12)
- discord/slack?
- When will the arxiv paper be released? HOT 4
- Criteria for evaluation HOT 1
- Train_Dataset Question
- Pre-prefix for the validation and test data? HOT 1
- SaTML Presentations
- Timeline for releasing final leaderboard HOT 2
- Can you offer a list of models trained on The Pile? HOT 7
- language-element.trad.character identified in baseline method HOT 6
- Restrictions on usage of model HOT 2
- calculating loss per token HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from lm-extraction-benchmark.