Comments (5)
Fourth decimal is 1e-4, this seems acceptable for such an old model given different settings / torch versions no?
from transformers.
I don't change any settings between the runs and what I ideally expect is the same scores, as the dataset and the model weights don't change. So I had a doubt if this variation is expected for rouge scores or not
from transformers.
Again for a very old model, this should be expected
from transformers.
Ok thank you for your response.
from transformers.
This issue has been automatically marked as stale because it has not had recent activity. If you think this still needs to be addressed please comment on this thread.
Please note that issues that do not follow the contributing guidelines are likely to be ignored.
from transformers.
Related Issues (20)
- Cannot convert llama 3 model to hf HOT 2
- error when using PPO in Gemma HOT 9
- Llama3 models causing `TypeError: not a string` error in LlamaTokenizer HOT 4
- Some functional problems in the implementation of Speculative Decoding HOT 3
- Error During Training with PatchTSMixerForTimeSeriesClassification for Time Series Classification HOT 1
- Whisper assistant decoding not working with pipeline
- Error During Training with PatchTSMixerForTimeSeriesClassification for Time Series Classification HOT 1
- TypeError: WhisperForConditionalGeneration.forward() got an unexpected keyword argument 'model' HOT 5
- FutureWarning about resume_download is raised after huggingface-hub 0.23.0 release
- Remove pipelines, chatformatters, templates etc --> Replace with simple generator function / manual string interpolation ---> Just have one standardized way for building datasets and running inference HOT 2
- HTML Files Keep on Loading HOT 1
- Wav2Vec2ForCTC weight mismatch HOT 1
- More memory consumption than litgpt
- Setting compute_metrics in Trainer with Idefics2ForConditionalGeneration leads to AttributeError: 'DynamicCache' object has no attribute 'detach' HOT 5
- DPT implementation contains unused parameters HOT 4
- Urdu Encoding Issue in Hugging Face Tokenizer HOT 1
- Add Prismatic VLMs to Transformers HOT 3
- Error converting from PyTorch to HuggingFace - Mistral / Mixtral
- model_max_length default parameters are missing in transformers>=4.40.0 HOT 2
- (Have PR) Speed up `BeamScorer` to make GPT-2 generation 2-3x faster HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from transformers.