{"payload":{"pageCount":2,"repositories":[{"type":"Public","name":"aug-pe","owner":"AI-secure","isFork":false,"description":"[ICML 2024] Differentially Private Synthetic Data via Foundation Model APIs 2: Text","allTopics":["language-model","differential-privacy","ai-privacy","large-language-models","prompt-engineering"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":18,"forksCount":2,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,3,5,0,0,0,0,0,0,0,0,0,0,0,0,0,1],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-21T12:33:02.757Z"}},{"type":"Public","name":"MMDT","owner":"AI-secure","isFork":false,"description":"Comprehensive Assessment of Trustworthiness in Multimodal Foundation Models","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-19T17:16:05.606Z"}},{"type":"Public","name":"DecodingTrust","owner":"AI-secure","isFork":false,"description":"A Comprehensive Assessment of Trustworthiness in GPT Models","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":10,"starsCount":223,"forksCount":50,"license":"Creative Commons Attribution Share Alike 4.0 International","participation":[1,0,1,0,0,83,1,0,6,14,3,6,2,5,46,83,68,50,11,4,10,0,0,10,1,3,11,6,1,1,4,0,0,2,1,13,0,0,0,0,0,0,2,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-19T01:24:30.178Z"}},{"type":"Public","name":"helm","owner":"AI-secure","isFork":true,"description":"Holistic Evaluation of Language Models (HELM), a framework to increase the transparency of language models (https://arxiv.org/abs/2211.09110).","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":2,"issueCount":0,"starsCount":0,"forksCount":232,"license":"Apache License 2.0","participation":[9,9,6,2,4,5,9,7,10,0,1,8,10,3,10,7,14,24,2,10,13,4,8,34,22,16,0,3,27,15,20,18,29,16,15,28,14,16,27,16,21,12,7,10,8,27,10,6,26,18,10,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-06-12T02:48:34.370Z"}},{"type":"Public","name":"VFL-ADMM","owner":"AI-secure","isFork":false,"description":"Improving Privacy-Preserving Vertical Federated Learning by Efficient Communication with ADMM (SaTML 2024)","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":"Apache License 2.0","participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-03-20T21:37:35.870Z"}},{"type":"Public","name":"DPFL-Robustness","owner":"AI-secure","isFork":false,"description":"[CCS 2023] Unraveling the Connections between Privacy and Certified Robustness in Federated Learning Against Poisoning Attacks ","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-02-15T06:49:21.030Z"}},{"type":"Public","name":"hf-blog","owner":"AI-secure","isFork":true,"description":"Public repo for HF blog posts","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":651,"license":null,"participation":[8,11,7,30,19,11,13,4,31,21,19,34,15,21,11,9,10,16,10,16,3,4,6,20,12,7,4,8,4,14,9,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2024-01-26T19:11:47.752Z"}},{"type":"Public","name":"DecodingTrust-Data-Legacy","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":[0,0,1,0,0,78,0,0,1,6,0,1,2,1,4,6,5,0,0,2,0,0,0,0,0,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-12-25T22:21:19.521Z"}},{"type":"Public","name":"TextGuard","owner":"AI-secure","isFork":false,"description":"TextGuard: Provable Defense against Backdoor Attacks on Text Classification","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1,0,1,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-11-07T07:17:16.937Z"}},{"type":"Public","name":"InfoBERT","owner":"AI-secure","isFork":false,"description":"[ICLR 2021] \"InfoBERT: Improving Robustness of Language Models from An Information Theoretic Perspective\" by Boxin Wang, Shuohang Wang, Yu Cheng, Zhe Gan, Ruoxi Jia, Bo Li, Jingjing Liu","allTopics":["information-theory","language-models","bert","adversarial-attacks","roberta","adversarial-defense","adversarial-robustness"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":82,"forksCount":7,"license":null,"participation":[0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,5,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0,0],"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-25T18:55:13.693Z"}},{"type":"Public","name":"FedGame","owner":"AI-secure","isFork":false,"description":"Official implementation for paper \"FedGame: A Game-Theoretic Defense against Backdoor Attacks in Federated Learning\" (NeurIPS 2023).","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":2,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-12T00:02:28.701Z"}},{"type":"Public","name":"Robustness-Against-Backdoor-Attacks","owner":"AI-secure","isFork":false,"description":"RAB: Provable Robustness Against Backdoor Attacks","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":5,"issueCount":2,"starsCount":39,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T21:46:41.474Z"}},{"type":"Public","name":"semantic-randomized-smoothing","owner":"AI-secure","isFork":false,"description":"[CCS 2021] TSS: Transformation-specific smoothing for robustness certification","allTopics":["security","robustness-verification","deep-learning"],"primaryLanguage":{"name":"Roff","color":"#ecdebe"},"pullRequestCount":5,"issueCount":0,"starsCount":20,"forksCount":3,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-10-03T21:45:55.203Z"}},{"type":"Public","name":"FLBenchmark-toolkit","owner":"AI-secure","isFork":false,"description":"Federated Learning Framework Benchmark (UniFed)","allTopics":["benchmark","federated-learning","machine-learning"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":5,"starsCount":46,"forksCount":5,"license":"Apache License 2.0","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-06-14T05:32:37.733Z"}},{"type":"Public","name":"SecretGen","owner":"AI-secure","isFork":false,"description":"A general model inversion attack against large pre-trained models.","allTopics":["machine-learning","privacy"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":2,"license":"MIT License","participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-22T20:37:17.413Z"}},{"type":"Public","name":"adversarial-glue","owner":"AI-secure","isFork":false,"description":"[NeurIPS 2021] \"Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of Language Models\" by Boxin Wang*, Chejian Xu*, Shuohang Wang, Zhe Gan, Yu Cheng, Jianfeng Gao, Ahmed Hassan Awadallah, Bo Li.","allTopics":["nlp","machine-learning","nlp-library","adversarial-examples","adversarial-attacks"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-04-03T08:06:14.880Z"}},{"type":"Public","name":"VeriGauge","owner":"AI-secure","isFork":false,"description":"A united toolbox for running major robustness verification approaches for DNNs. [S&P 2023]","allTopics":["robustness","verfication","deep-learning"],"primaryLanguage":{"name":"C","color":"#555555"},"pullRequestCount":4,"issueCount":3,"starsCount":86,"forksCount":6,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-03-24T22:24:33.933Z"}},{"type":"Public","name":"Certified-Fairness","owner":"AI-secure","isFork":false,"description":"[NeurIPS 2022] Code for Certifying Some Distributional Fairness with Subpopulation Decomposition","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":4,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2023-01-03T06:12:19.470Z"}},{"type":"Public","name":"CoPur","owner":"AI-secure","isFork":false,"description":"CoPur: Certifiably Robust Collaborative Inference via Feature Purification (NeurIPS 2022)","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":7,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-07T04:37:17.676Z"}},{"type":"Public","name":"transferability-versus-robustness","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-06T17:08:23.768Z"}},{"type":"Public","name":"DMLW2022","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"HTML","color":"#e34c26"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-12-03T21:24:40.807Z"}},{"type":"Public","name":"Certified-Robustness-SoK-Oldver","owner":"AI-secure","isFork":false,"description":"This repo keeps track of popular provable training and verification approaches towards robust neural networks, including leaderboards on popular datasets and paper categorization.","allTopics":[],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":99,"forksCount":10,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-18T21:04:43.723Z"}},{"type":"Public","name":"Layerwise-Orthogonal-Training","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":5,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-10-11T20:30:47.862Z"}},{"type":"Public","name":"CROP","owner":"AI-secure","isFork":false,"description":"[ICLR 2022] CROP: Certifying Robust Policies for Reinforcement Learning through Functional Smoothing","allTopics":["reinforcement-learning","certification","robustness"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":1,"starsCount":8,"forksCount":2,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-16T00:09:44.508Z"}},{"type":"Public","name":"SemAttack","owner":"AI-secure","isFork":false,"description":"[NAACL 2022] \"SemAttack: Natural Textual Attacks via Different Semantic Spaces\" by Boxin Wang, Chejian Xu, Xiangyu Liu, Yu Cheng, Bo Li","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":19,"forksCount":5,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-06-11T10:17:59.532Z"}},{"type":"Public","name":"Transferability-Reduced-Smooth-Ensemble","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":3,"starsCount":22,"forksCount":8,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-15T18:01:56.293Z"}},{"type":"Public","name":"COPA","owner":"AI-secure","isFork":false,"description":"[ICLR 2022] COPA: Certifying Robust Policies for Offline Reinforcement Learning against Poisoning Attacks","allTopics":["certification","robustness","reinforcement-learning"],"primaryLanguage":null,"pullRequestCount":0,"issueCount":0,"starsCount":8,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-04-05T23:31:03.919Z"}},{"type":"Public","name":"COPA_Highway","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Jupyter Notebook","color":"#DA5B0B"},"pullRequestCount":0,"issueCount":0,"starsCount":0,"forksCount":0,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-19T06:44:00.620Z"}},{"type":"Public","name":"COPA_Atari","owner":"AI-secure","isFork":false,"description":"","allTopics":[],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":1,"forksCount":1,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2022-03-18T00:39:46.554Z"}},{"type":"Public","name":"DataLens","owner":"AI-secure","isFork":false,"description":"[CCS 2021] \"DataLens: Scalable Privacy Preserving Training via Gradient Compression and Aggregation\" by Boxin Wang*, Fan Wu*, Yunhui Long*, Luka Rimanic, Ce Zhang, Bo Li","allTopics":["generative-model","differential-privacy","gradient-compression"],"primaryLanguage":{"name":"Python","color":"#3572A5"},"pullRequestCount":0,"issueCount":0,"starsCount":37,"forksCount":7,"license":null,"participation":null,"lastUpdated":{"hasBeenPushedTo":true,"timestamp":"2021-12-28T20:00:32.484Z"}}],"repositoryCount":52,"userInfo":null,"searchable":true,"definitions":[],"typeFilters":[{"id":"all","text":"All"},{"id":"public","text":"Public"},{"id":"source","text":"Sources"},{"id":"fork","text":"Forks"},{"id":"archived","text":"Archived"},{"id":"template","text":"Templates"}],"compactMode":false},"title":"Repositories"}