Skip to content

Commit

Permalink
Automated report
Browse files Browse the repository at this point in the history
  • Loading branch information
deep-diver committed Jan 30, 2024
1 parent efdaea8 commit cc1c656
Show file tree
Hide file tree
Showing 20 changed files with 32 additions and 0 deletions.
8 changes: 8 additions & 0 deletions tags/Computer Vision.md
Original file line number Diff line number Diff line change
Expand Up @@ -75,3 +75,11 @@
- [From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+From+GPT-4+to+Gemini+and+Beyond%3A+Assessing+the+Landscape+of+MLLMs+on+Generalizability%2C+Trustworthiness+and+Causality+through+Four+Modalities.yaml) / 2024-01-29
- [TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+TIP-Editor%3A+An+Accurate+3D+Editor+Following+Both+Text-Prompts+And+Image-Prompts.yaml) / 2024-01-29
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Divide and Conquer: Language Models can Plan and Self-Correct for Compositional Text-to-Image Generation](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Divide+and+Conquer%3A+Language+Models+can+Plan+and+Self-Correct+for+Compositional+Text-to-Image+Generation.yaml) / 2024-01-30
- [InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+InternLM-XComposer2%3A+Mastering+Free-form+Text-Image+Composition+and+Comprehension+in+Vision-Language+Large+Model.yaml) / 2024-01-30
- [Media2Face: Co-speech Facial Animation Generation With Multi-Modality Guidance](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Media2Face%3A+Co-speech+Facial+Animation+Generation+With+Multi-Modality+Guidance.yaml) / 2024-01-30
- [Mobile-Agent: Autonomous Multi-Modal Mobile Device Agent with Visual Perception](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Mobile-Agent%3A+Autonomous+Multi-Modal+Mobile+Device+Agent+with+Visual+Perception.yaml) / 2024-01-30
- [Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Motion-I2V%3A+Consistent+and+Controllable+Image-to-Video+Generation+with+Explicit+Motion+Modeling.yaml) / 2024-01-30
- [Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Object-Driven+One-Shot+Fine-tuning+of+Text-to-Image+Diffusion+with+Prototypical+Embedding.yaml) / 2024-01-30
- [Overcoming the Pitfalls of Vision-Language Model Finetuning for OOD Generalization](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Overcoming+the+Pitfalls+of+Vision-Language+Model+Finetuning+for+OOD+Generalization.yaml) / 2024-01-30
- [StableIdentity: Inserting Anybody into Anywhere at First Sight](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+StableIdentity%3A+Inserting+Anybody+into+Anywhere+at+First+Sight.yaml) / 2024-01-30
8 changes: 8 additions & 0 deletions tags/Deep Learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -108,3 +108,11 @@
- [SliceGPT: Compress Large Language Models by Deleting Rows and Columns](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+SliceGPT%3A+Compress+Large+Language+Models+by+Deleting+Rows+and+Columns.yaml) / 2024-01-29
- [TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+TIP-Editor%3A+An+Accurate+3D+Editor+Following+Both+Text-Prompts+And+Image-Prompts.yaml) / 2024-01-29
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Divide and Conquer: Language Models can Plan and Self-Correct for Compositional Text-to-Image Generation](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Divide+and+Conquer%3A+Language+Models+can+Plan+and+Self-Correct+for+Compositional+Text-to-Image+Generation.yaml) / 2024-01-30
- [InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+InternLM-XComposer2%3A+Mastering+Free-form+Text-Image+Composition+and+Comprehension+in+Vision-Language+Large+Model.yaml) / 2024-01-30
- [Media2Face: Co-speech Facial Animation Generation With Multi-Modality Guidance](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Media2Face%3A+Co-speech+Facial+Animation+Generation+With+Multi-Modality+Guidance.yaml) / 2024-01-30
- [MoE-LLaVA: Mixture of Experts for Large Vision-Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+MoE-LLaVA%3A+Mixture+of+Experts+for+Large+Vision-Language+Models.yaml) / 2024-01-30
- [Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Motion-I2V%3A+Consistent+and+Controllable+Image-to-Video+Generation+with+Explicit+Motion+Modeling.yaml) / 2024-01-30
- [Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Object-Driven+One-Shot+Fine-tuning+of+Text-to-Image+Diffusion+with+Prototypical+Embedding.yaml) / 2024-01-30
- [Overcoming the Pitfalls of Vision-Language Model Finetuning for OOD Generalization](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Overcoming+the+Pitfalls+of+Vision-Language+Model+Finetuning+for+OOD+Generalization.yaml) / 2024-01-30
- [StableIdentity: Inserting Anybody into Anywhere at First Sight](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+StableIdentity%3A+Inserting+Anybody+into+Anywhere+at+First+Sight.yaml) / 2024-01-30
2 changes: 2 additions & 0 deletions tags/Emerging Applications of Machine Learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -14,3 +14,5 @@
- [CreativeSynth: Creative Blending and Synthesis of Visual Arts based on Multimodal Diffusion](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/11/2024-01-26+CreativeSynth%3A+Creative+Blending+and+Synthesis+of+Visual+Arts+based+on+Multimodal+Diffusion.yaml) / 2024-01-26
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/12/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Object-Driven+One-Shot+Fine-tuning+of+Text-to-Image+Diffusion+with+Prototypical+Embedding.yaml) / 2024-01-30
- [StableIdentity: Inserting Anybody into Anywhere at First Sight](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+StableIdentity%3A+Inserting+Anybody+into+Anywhere+at+First+Sight.yaml) / 2024-01-30
1 change: 1 addition & 0 deletions tags/Explainable AI and Interpretability.md
Original file line number Diff line number Diff line change
Expand Up @@ -16,3 +16,4 @@
- [EAGLE: Speculative Sampling Requires Rethinking Feature Uncertainty](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/12/2024-01-29+EAGLE%3A+Speculative+Sampling+Requires+Rethinking+Feature+Uncertainty.yaml) / 2024-01-29
- [From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/12/2024-01-29+From+GPT-4+to+Gemini+and+Beyond%3A+Assessing+the+Landscape+of+MLLMs+on+Generalizability%2C+Trustworthiness+and+Causality+through+Four+Modalities.yaml) / 2024-01-29
- [From GPT-4 to Gemini and Beyond: Assessing the Landscape of MLLMs on Generalizability, Trustworthiness and Causality through Four Modalities](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+From+GPT-4+to+Gemini+and+Beyond%3A+Assessing+the+Landscape+of+MLLMs+on+Generalizability%2C+Trustworthiness+and+Causality+through+Four+Modalities.yaml) / 2024-01-29
- [Motion-I2V: Consistent and Controllable Image-to-Video Generation with Explicit Motion Modeling](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Motion-I2V%3A+Consistent+and+Controllable+Image-to-Video+Generation+with+Explicit+Motion+Modeling.yaml) / 2024-01-30
Original file line number Diff line number Diff line change
@@ -1,2 +1,3 @@
- [Towards Conversational Diagnostic AI](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/3/2024-01-12+Towards+Conversational+Diagnostic+AI.yaml) / 2024-01-12
- [Rambler: Supporting Writing With Speech via LLM-Assisted Gist Manipulation](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/7/2024-01-22+Rambler%3A+Supporting+Writing+With+Speech+via+LLM-Assisted+Gist+Manipulation.yaml) / 2024-01-22
- [Media2Face: Co-speech Facial Animation Generation With Multi-Modality Guidance](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Media2Face%3A+Co-speech+Facial+Animation+Generation+With+Multi-Modality+Guidance.yaml) / 2024-01-30
5 changes: 5 additions & 0 deletions tags/Natural Language Processing.md
Original file line number Diff line number Diff line change
Expand Up @@ -85,3 +85,8 @@
- [Generative Expressive Robot Behaviors using Large Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Generative+Expressive+Robot+Behaviors+using+Large+Language+Models.yaml) / 2024-01-29
- [TIP-Editor: An Accurate 3D Editor Following Both Text-Prompts And Image-Prompts](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+TIP-Editor%3A+An+Accurate+3D+Editor+Following+Both+Text-Prompts+And+Image-Prompts.yaml) / 2024-01-29
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Divide and Conquer: Language Models can Plan and Self-Correct for Compositional Text-to-Image Generation](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Divide+and+Conquer%3A+Language+Models+can+Plan+and+Self-Correct+for+Compositional+Text-to-Image+Generation.yaml) / 2024-01-30
- [InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+InternLM-XComposer2%3A+Mastering+Free-form+Text-Image+Composition+and+Comprehension+in+Vision-Language+Large+Model.yaml) / 2024-01-30
- [Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Object-Driven+One-Shot+Fine-tuning+of+Text-to-Image+Diffusion+with+Prototypical+Embedding.yaml) / 2024-01-30
- [Overcoming the Pitfalls of Vision-Language Model Finetuning for OOD Generalization](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Overcoming+the+Pitfalls+of+Vision-Language+Model+Finetuning+for+OOD+Generalization.yaml) / 2024-01-30
- [Rephrasing the Web: A Recipe for Compute and Data-Efficient Language Modeling](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Rephrasing+the+Web%3A+A+Recipe+for+Compute+and+Data-Efficient+Language+Modeling.yaml) / 2024-01-30
1 change: 1 addition & 0 deletions tags/Reinforcement Learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,3 +8,4 @@
- [Self-Rewarding Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/6/2024-01-19+Self-Rewarding+Language+Models.yaml) / 2024-01-19
- [WARM: On the Benefits of Weight Averaged Reward Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/8/2024-01-23+WARM%3A+On+the+Benefits+of+Weight+Averaged+Reward+Models.yaml) / 2024-01-23
- [Large-scale Reinforcement Learning for Diffusion Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/9/2024-01-24+Large-scale+Reinforcement+Learning+for+Diffusion+Models.yaml) / 2024-01-24
- [SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+SERL%3A+A+Software+Suite+for+Sample-Efficient+Robotic+Reinforcement+Learning.yaml) / 2024-01-30
1 change: 1 addition & 0 deletions tags/Robotics and Control.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,3 +5,4 @@
- [Adaptive Mobile Manipulation for Articulated Objects In the Open World](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/11/2024-01-26+Adaptive+Mobile+Manipulation+for+Articulated+Objects+In+the+Open+World.yaml) / 2024-01-26
- [Generative Expressive Robot Behaviors using Large Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/12/2024-01-29+Generative+Expressive+Robot+Behaviors+using+Large+Language+Models.yaml) / 2024-01-29
- [Generative Expressive Robot Behaviors using Large Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Generative+Expressive+Robot+Behaviors+using+Large+Language+Models.yaml) / 2024-01-29
- [SERL: A Software Suite for Sample-Efficient Robotic Reinforcement Learning](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+SERL%3A+A+Software+Suite+for+Sample-Efficient+Robotic+Reinforcement+Learning.yaml) / 2024-01-30
5 changes: 5 additions & 0 deletions tags/Supervised Learning.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,3 +37,8 @@
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/12/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [Learning Universal Predictors](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Learning+Universal+Predictors.yaml) / 2024-01-29
- [Taiyi-Diffusion-XL: Advancing Bilingual Text-to-Image Generation with Large Vision-Language Model Support](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/13/2024-01-29+Taiyi-Diffusion-XL%3A+Advancing+Bilingual+Text-to-Image+Generation+with+Large+Vision-Language+Model+Support.yaml) / 2024-01-29
- [InternLM-XComposer2: Mastering Free-form Text-Image Composition and Comprehension in Vision-Language Large Model](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+InternLM-XComposer2%3A+Mastering+Free-form+Text-Image+Composition+and+Comprehension+in+Vision-Language+Large+Model.yaml) / 2024-01-30
- [MoE-LLaVA: Mixture of Experts for Large Vision-Language Models](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+MoE-LLaVA%3A+Mixture+of+Experts+for+Large+Vision-Language+Models.yaml) / 2024-01-30
- [Object-Driven One-Shot Fine-tuning of Text-to-Image Diffusion with Prototypical Embedding](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Object-Driven+One-Shot+Fine-tuning+of+Text-to-Image+Diffusion+with+Prototypical+Embedding.yaml) / 2024-01-30
- [Overcoming the Pitfalls of Vision-Language Model Finetuning for OOD Generalization](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Overcoming+the+Pitfalls+of+Vision-Language+Model+Finetuning+for+OOD+Generalization.yaml) / 2024-01-30
- [Rephrasing the Web: A Recipe for Compute and Data-Efficient Language Modeling](https://github.com/deep-diver/hf-daily-paper-newsletter/blob/main/archive/14/2024-01-30+Rephrasing+the+Web%3A+A+Recipe+for+Compute+and+Data-Efficient+Language+Modeling.yaml) / 2024-01-30

0 comments on commit cc1c656

Please sign in to comment.