{"payload":{"pageCount":1,"repositories":[{"type":"Public","name":"llm-adaptive-attacks","owner":"tml-epfl","isFork":false,"description":"Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks [arXiv, Apr 2024]","allTopics":[],"primaryLanguage":{"name":"Shell","color":"#89e051"},"pullRequestCount":0,"issueCount":0,"starsCount":183,"forksCount":20,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-09-20T14:53:12.309Z"}},{"type":"Public","name":"llm-past-tense","owner":"tml-epfl","isFork":false,"description":"Does Refusal Training in LLMs Generalize to the Past Tense? [arXiv, July 2024]","allTopics":["jailbreaking","robustness","generalization","llms"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":50,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-07-18T23:02:55.251Z"}},{"type":"Public","name":"icl-alignment","owner":"tml-epfl","isFork":false,"description":"Is In-Context Learning Sufficient for Instruction Following in LLMs?","allTopics":["alignment","instruction-following","in-context-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":20,"forksCount":3,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-31T07:34:10.210Z"}},{"type":"Public","name":"long-is-more-for-alignment","owner":"tml-epfl","isFork":false,"description":"Long Is More for Alignment: A Simple but Tough-to-Beat Baseline for Instruction Fine-Tuning [ICML 2024]","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":12,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-05-02T07:34:18.282Z"}},{"type":"Public","name":"why-weight-decay","owner":"tml-epfl","isFork":false,"description":"Why Do We Need Weight Decay in Modern Deep Learning? [arXiv, Oct 2023]","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":41,"forksCount":0,"license":"Other","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-09T09:45:56.110Z"}},{"type":"Public","name":"sam-low-rank-features","owner":"tml-epfl","isFork":false,"description":"Sharpness-Aware Minimization Leads to Low-Rank Features [NeurIPS 2023]","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":1,"starsCount":24,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-22T08:56:28.653Z"}},{"type":"Public","name":"sharpness-vs-generalization","owner":"tml-epfl","isFork":false,"description":"A modern look at the relationship between sharpness and generalization [ICML 2023]","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":42,"forksCount":3,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-09-11T10:33:06.548Z"}},{"type":"Public","name":"sgd-sparse-features","owner":"tml-epfl","isFork":false,"description":"SGD with large step sizes learns sparse features [ICML 2023]","allTopics":["sgd","implicit-bias","generalization","large-step-sizes"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":31,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-24T22:07:18.130Z"}},{"type":"Public","name":"tml-epfl.github.io","owner":"tml-epfl","isFork":false,"description":"Creating a repository to store all related information for the weekly TML group meetings. ","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"MIT License","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-11-16T13:39:55.704Z"}},{"type":"Public","name":"understanding-sam","owner":"tml-epfl","isFork":false,"description":"Towards Understanding Sharpness-Aware Minimization [ICML 2022]","allTopics":["generalization","sharpness","flatness","sharpness-aware-minimization","understanding-deep-learning"],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":34,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-14T01:45:26.183Z"}},{"type":"Public","name":"adv-training-corruptions","owner":"tml-epfl","isFork":false,"description":"On the effectiveness of adversarial training against common corruptions [UAI 2022]","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":30,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-05-16T07:38:18.040Z"}},{"type":"Public","name":"understanding-fast-adv-training","owner":"tml-epfl","isFork":false,"description":"Understanding and Improving Fast Adversarial Training [NeurIPS 2020]","allTopics":["robust-optimization","robustness","adversarial-examples","adversarial-training"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":94,"forksCount":12,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-09-23T07:43:43.627Z"}}],"repositoryCount":12,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"tml-epfl repositories"}