-
-
Notifications
You must be signed in to change notification settings - Fork 15
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Different ranges in the results graphs #991
Comments
👋 Hello @Josephts1, thank you for raising this issue about Ultralytics HUB 🚀! It looks like you’re working on training YOLO models for mandarin detection, and we appreciate you sharing your training details along with the results graphs 📊. To help address your question, please take a look at our HUB Docs for guidance on best practices and potential insights:
If you're observing varying loss graph ranges, this could depend on the model's architecture, training dynamics, or specific dataset characteristics (scale, variability, annotations). Please share additional context, such as:
If this is a potential 🐛 Bug Report, please also include a minimum reproducible example (MRE) to assist us in reproducing the behavior. Our engineering team will look into this further and get back to you soon. Your patience and detailed input are much appreciated, as they help us continue improving the HUB platform! 🚀😊 |
@Josephts1 thank you for your detailed explanation and the accompanying graphs! It’s great to see your experimentation with different training strategies for mandarin detection. The variation in loss ranges you observed is a common occurrence and can be explained by the following factors:
Recommendations:
If you'd like to further analyze or adjust training behavior, consider visualizing additional metrics or leveraging tools provided in the Ultralytics HUB. The HUB allows for streamlined dataset management, training, and monitoring of results. Let me know if you have further questions or need clarification! 😊 |
Search before asking
Question
I'm looking for the best way to train a YOLO model for mandarin detection with my own dataset.
First, I tried with a pre-trained model (model=YOLO('yolo11s.pt')). I got fairly good results but there were a lot of peaks and troughs in the graphs (see image 1). The execution time was too long.
Second, I tried with the same pre-trained model but freezing the 24 layers of the YOLO model. I got better results (better mAP50-95 and better convergence) but there were still peaks and troughs (see image 2).
Finally, I trained a model from scratch (model=YOLO('yolo11s.yaml')) and got the best result so far (see image 3).
My question is: If you look at the vertical axis of the train/box_loss, val/box_loss, train/cls_loss, val/cls_loss, train/dfl_loss and val/dfl_loss graphs, they vary a lot between the pre-trained model and the new model. Does anyone know why these ranges are so different, one has values between 1 and 0, the other has values between 5 and 0?
(imagen 1)
![Image](https://private-user-images.githubusercontent.com/89653047/405790708-d7963708-641e-482a-abec-88afd5f6a16b.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzg5ODcyNzMsIm5iZiI6MTczODk4Njk3MywicGF0aCI6Ii84OTY1MzA0Ny80MDU3OTA3MDgtZDc5NjM3MDgtNjQxZS00ODJhLWFiZWMtODhhZmQ1ZjZhMTZiLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDglMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA4VDAzNTYxM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPThkYTRkZDkyODdiN2NlMDM3ZjFkN2EwODhkNjA4OTcxMDA2NzcxZGIxNWJkZGM4NTM5MjA0Zjk0M2M5ZTc1YjgmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.xv2Ad4O1oFN-9tADSw05ybmldeWwi11stCtqcZsiqdw)
![Image](https://private-user-images.githubusercontent.com/89653047/405791418-7f7ff16a-6940-4bf4-bff1-c318f416edfc.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzg5ODcyNzMsIm5iZiI6MTczODk4Njk3MywicGF0aCI6Ii84OTY1MzA0Ny80MDU3OTE0MTgtN2Y3ZmYxNmEtNjk0MC00YmY0LWJmZjEtYzMxOGY0MTZlZGZjLnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDglMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA4VDAzNTYxM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWIxNGMyN2UwYjdlN2RiNTBjYWE4MTk5ODA2M2VlZmE5MDZhYjBjMjI5NjQyYTliNGRhNzljMzdmY2ZlOTIxM2MmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.XYU_1ysEtHbUdmqiezep9Pw7eF7JMjMi64p41oQ4UeM)
![Image](https://private-user-images.githubusercontent.com/89653047/405792792-65408b98-3776-42a1-b8cc-9e512d53e367.png?jwt=eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJpc3MiOiJnaXRodWIuY29tIiwiYXVkIjoicmF3LmdpdGh1YnVzZXJjb250ZW50LmNvbSIsImtleSI6ImtleTUiLCJleHAiOjE3Mzg5ODcyNzMsIm5iZiI6MTczODk4Njk3MywicGF0aCI6Ii84OTY1MzA0Ny80MDU3OTI3OTItNjU0MDhiOTgtMzc3Ni00MmExLWI4Y2MtOWU1MTJkNTNlMzY3LnBuZz9YLUFtei1BbGdvcml0aG09QVdTNC1ITUFDLVNIQTI1NiZYLUFtei1DcmVkZW50aWFsPUFLSUFWQ09EWUxTQTUzUFFLNFpBJTJGMjAyNTAyMDglMkZ1cy1lYXN0LTElMkZzMyUyRmF3czRfcmVxdWVzdCZYLUFtei1EYXRlPTIwMjUwMjA4VDAzNTYxM1omWC1BbXotRXhwaXJlcz0zMDAmWC1BbXotU2lnbmF0dXJlPWFiYmViMjA2NTc0Y2UxMDlkY2I4ZWRhZmUyZWJjODQ0YTI5NzcwYWVjN2YwNTgyOTFmZmRjYTMyMmFhZGEyMzgmWC1BbXotU2lnbmVkSGVhZGVycz1ob3N0In0.EMmfCigOhkhpgzrkDrgggWqmTfrXspD5h7XucbCzxCc)
(imagen 2)
(imagen 3)
Additional
No response
The text was updated successfully, but these errors were encountered: