{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":550048842,"defaultBranch":"united","name":"KoboldAI","ownerLogin":"pi6am","currentUserCanPush":false,"isFork":true,"isEmpty":false,"createdAt":"2022-10-12T06:01:31.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/34464159?v=4","public":true,"private":false,"isOrgOwned":false},"refInfo":{"name":"","listCacheKey":"v0:1693376066.0","currentOid":""},"activityList":{"items":[{"before":"40e258dcda3a71a9fdbad30f417d182af0da798c","after":"d13f3a0d3bde2c33369385c8f000cf8dd97b9841","ref":"refs/heads/united","pushedAt":"2024-07-27T04:24:19.000Z","pushType":"pr_merge","commitsCount":43,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #42 from henk717/united\n\nMerge upstream","shortMessageHtmlLink":"Merge pull request #42 from henk717/united"}},{"before":"7a43ac32c366df12303f7a8636fb53340725f8ac","after":"40e258dcda3a71a9fdbad30f417d182af0da798c","ref":"refs/heads/united","pushedAt":"2024-01-31T05:52:51.000Z","pushType":"pr_merge","commitsCount":44,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #41 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #41 from henk717/united"}},{"before":"a4ff8a5285d7a6f288a39c99eaf62a2922bb6f57","after":"7a43ac32c366df12303f7a8636fb53340725f8ac","ref":"refs/heads/united","pushedAt":"2024-01-01T21:08:58.000Z","pushType":"pr_merge","commitsCount":63,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #40 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #40 from henk717/united"}},{"before":"b20337a3ae690e0ee585f923ecc5cf0d8a0652d0","after":"a4ff8a5285d7a6f288a39c99eaf62a2922bb6f57","ref":"refs/heads/united","pushedAt":"2023-10-25T03:33:11.000Z","pushType":"pr_merge","commitsCount":26,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #39 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #39 from henk717/united"}},{"before":"3a977578e622e83555fab755c1a87d779132e0d9","after":"b20337a3ae690e0ee585f923ecc5cf0d8a0652d0","ref":"refs/heads/united","pushedAt":"2023-10-05T07:35:05.000Z","pushType":"pr_merge","commitsCount":12,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #38 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #38 from henk717/united"}},{"before":"a1b82bdab10d4e23e9375e3fce2d931685e25b2a","after":"3a977578e622e83555fab755c1a87d779132e0d9","ref":"refs/heads/united","pushedAt":"2023-09-26T04:22:53.000Z","pushType":"pr_merge","commitsCount":9,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #37 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #37 from henk717/united"}},{"before":"fc538b00aa35e466879f202845e4b47d20ebf65a","after":"a1b82bdab10d4e23e9375e3fce2d931685e25b2a","ref":"refs/heads/united","pushedAt":"2023-09-17T23:01:30.000Z","pushType":"pr_merge","commitsCount":8,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #36 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #36 from henk717/united"}},{"before":"71181fc1fef148916ef5d0e3fc7d76436b1f4ee1","after":"fc538b00aa35e466879f202845e4b47d20ebf65a","ref":"refs/heads/united","pushedAt":"2023-09-13T06:59:28.000Z","pushType":"pr_merge","commitsCount":7,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #35 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #35 from henk717/united"}},{"before":"dda5acd5d5697d0bb216f8cee7324c8965fb3549","after":"71181fc1fef148916ef5d0e3fc7d76436b1f4ee1","ref":"refs/heads/united","pushedAt":"2023-09-10T17:36:21.000Z","pushType":"pr_merge","commitsCount":133,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #34 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #34 from henk717/united"}},{"before":null,"after":"d6ed75f9938e81072362856ad2649a20aa5f59a5","ref":"refs/heads/feat/exllama-unban-eos","pushedAt":"2023-08-30T06:14:26.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Hook up use_default_badwordids in exllama\n\nUse the value of the use_default_badwordids setting to configure\nbad_words_ids. Also add square brackets to bad_words_ids if the\nuse_default_badwordids setting is True. Fix an issue with\nattempting to use the tokenizer too early, and fix an exception\npopulating Lua bridge data when zero tokens are generated, which\ncan now happen if use_default_badwordids is False and the first\ntoken generated is EOS.","shortMessageHtmlLink":"Hook up use_default_badwordids in exllama"}},{"before":"2c48e05f7c2eb053e7a2d196f3b24fd5335492ab","after":"b5b0e3faeae042e8d28b4965543a76fd715b5fb2","ref":"refs/heads/merge/united-exllama","pushedAt":"2023-08-30T02:51:55.000Z","pushType":"push","commitsCount":58,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge branch 'henk717/united' into merge/united-exllama","shortMessageHtmlLink":"Merge branch 'henk717/united' into merge/united-exllama"}},{"before":"6151cbd05303b10d631e99e523cba5744c2faf7e","after":"2c48e05f7c2eb053e7a2d196f3b24fd5335492ab","ref":"refs/heads/merge/united-exllama","pushedAt":"2023-08-28T16:55:38.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Add exllama dependency back to requirements.","shortMessageHtmlLink":"Add exllama dependency back to requirements."}},{"before":null,"after":"6151cbd05303b10d631e99e523cba5744c2faf7e","ref":"refs/heads/merge/united-exllama","pushedAt":"2023-08-28T16:35:19.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge branch 'united' into merge/united-exllama","shortMessageHtmlLink":"Merge branch 'united' into merge/united-exllama"}},{"before":null,"after":"554af7b1754fa2e574fbbcfa2a612b13969bda63","ref":"refs/heads/feat/exllama-config","pushedAt":"2023-08-28T07:08:51.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Modify exllama to load unrenamed gptq quantized models\n\nRead config.json and enable exllama loading if the model has a\n`quantization_config` with `quant_methdod` of `gptq`. Note that this\nimplementation is limited and only supports model.safetensors.\nThat said, this supports loading popular gptq quantized models\nwithout renaming or symlinking the model file.","shortMessageHtmlLink":"Modify exllama to load unrenamed gptq quantized models"}},{"before":null,"after":"08ff7c138c35e344819acbd82fa18e88732e08a4","ref":"refs/heads/feat/exllama-badwords","pushedAt":"2023-08-27T23:42:16.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Add the eos token to exllama bad words.\n\nThe bos token was already hardcoded as a bad word id.\nStore badwords in a list and iterate over them during generation.\nAdd the Llama eos token to the list of bad words.\nAlso support \"single line mode\", which adds newline (13) to badwords.","shortMessageHtmlLink":"Add the eos token to exllama bad words."}},{"before":"b7e38b47570cb910d4b5b9c853985e6d3fba9107","after":null,"ref":"refs/heads/fix/multinomial-workaround","pushedAt":"2023-08-27T05:42:28.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"}},{"before":null,"after":"b7e38b47570cb910d4b5b9c853985e6d3fba9107","ref":"refs/heads/fix/multinomial-workaround","pushedAt":"2023-08-27T05:38:28.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Resample to work around a bug in torch.multinomial\n\nThere is a bug in PyTorch 2.0.1 that allows torch.multinomial to\nsometimes choose elements that have zero probability. Since\nthis is uncommon we can continue to use torch.multinomial as\nlong as we verify that the results are valid. If they aren't,\ntry again until the probability of each selected token is positive.","shortMessageHtmlLink":"Resample to work around a bug in torch.multinomial"}},{"before":null,"after":"b96d5d8646b320096f06fc65a291469d8ca9a5dd","ref":"refs/heads/feat/exllama-stoppers","pushedAt":"2023-08-23T06:12:38.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Add stopper hooks suppport to exllama","shortMessageHtmlLink":"Add stopper hooks suppport to exllama"}},{"before":null,"after":"070cfd339a27ee1ef62fd74a495fb34b80a4920c","ref":"refs/heads/fix/exllama-eos-space","pushedAt":"2023-08-20T00:49:44.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Strip the eos token from exllama generations.\n\nThe end-of-sequence () token indicates the end of a generation.\nWhen a token sequence containing is decoded, an extra (wrong)\nspace is inserted at the beginning of the generation. To avoid this,\nstrip the eos token out of the result before returning it.\nThe eos token was getting stripped later, so this doesn't change\nthe output except to avoid the spurious leading space.","shortMessageHtmlLink":"Strip the eos token from exllama generations."}},{"before":"d8d9890f460cf269f897f9262755cb2d0e040f0d","after":"dda5acd5d5697d0bb216f8cee7324c8965fb3549","ref":"refs/heads/united","pushedAt":"2023-08-18T20:19:24.000Z","pushType":"pr_merge","commitsCount":4,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #33 from henk717/united\n\nMerge united.","shortMessageHtmlLink":"Merge pull request #33 from henk717/united"}},{"before":"01c32621ca19ea618dd6536980b5f568495c089d","after":"d8d9890f460cf269f897f9262755cb2d0e040f0d","ref":"refs/heads/united","pushedAt":"2023-08-13T21:10:35.000Z","pushType":"pr_merge","commitsCount":565,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #32 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #32 from henk717/united"}},{"before":"507da6fcf7af410018c70a01946122d5b77811df","after":"01c32621ca19ea618dd6536980b5f568495c089d","ref":"refs/heads/united","pushedAt":"2023-05-04T05:04:31.000Z","pushType":"pr_merge","commitsCount":12,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #31 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #31 from henk717/united"}},{"before":"376884854869528e169ccc7444b308cb41d1ae29","after":"35d344b9518b3c51efbb447784e626c280b55a9e","ref":"refs/heads/fix/llama-tokens","pushedAt":"2023-05-03T16:51:39.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Remove torch dependency and more generic dim0 workaround\n\nRemove torch dependency from hf.py\nMake workaround for dimension zero values of token_ids\nmore generic to handle every token, not just newlines.","shortMessageHtmlLink":"Remove torch dependency and more generic dim0 workaround"}},{"before":null,"after":"376884854869528e169ccc7444b308cb41d1ae29","ref":"refs/heads/fix/llama-tokens","pushedAt":"2023-05-03T08:29:27.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Fix tokenization and whitespace issues with llama-derived models\n\nWork around the 'soft' prefix space behavior of sentencepiece.\nOverride encode to restore the deleted HF support for decode_with_prefix_space.\nOverride decode to skip the soft space and return true decoded tokens.\nAllow submitting chat messages with embedded newlines.\nSplit sentences between punctuation and whitespace, rather than after whitespace.\nAlso include trailing quotes and brackets after sentence stoppers.\nThis avoids splitting .\" and .) into two tokens, for instance.\nInsert whitespace at the beginning of the author's note, since sentences are\nsplit with leading whitespace.\nRemove spurious newlines at the end of chat responses.","shortMessageHtmlLink":"Fix tokenization and whitespace issues with llama-derived models"}},{"before":"eb4e89c2fa2ab27831072ad7f33b5c1380b0f382","after":"507da6fcf7af410018c70a01946122d5b77811df","ref":"refs/heads/united","pushedAt":"2023-05-03T04:25:47.000Z","pushType":"pr_merge","commitsCount":142,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #30 from henk717/united\n\nMerge large refactor from united.","shortMessageHtmlLink":"Merge pull request #30 from henk717/united"}},{"before":"5189af2f2e24a54cfe62eba457c8874394a22edf","after":"eb4e89c2fa2ab27831072ad7f33b5c1380b0f382","ref":"refs/heads/united","pushedAt":"2023-04-30T21:20:12.000Z","pushType":"pr_merge","commitsCount":39,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #29 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #29 from henk717/united"}},{"before":"157b1c75e7f6317129a95ee7c3391cce1bcd20f9","after":"5189af2f2e24a54cfe62eba457c8874394a22edf","ref":"refs/heads/united","pushedAt":"2023-04-26T17:10:10.000Z","pushType":"pr_merge","commitsCount":46,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #28 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #28 from henk717/united"}},{"before":"5b8db52abb23024db0d50f308057c6df39481903","after":"157b1c75e7f6317129a95ee7c3391cce1bcd20f9","ref":"refs/heads/united","pushedAt":"2023-03-27T06:25:36.000Z","pushType":"pr_merge","commitsCount":6,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #27 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #27 from henk717/united"}},{"before":"835cf50ebc381d37367336d94bce532e1cb0b15e","after":"5b8db52abb23024db0d50f308057c6df39481903","ref":"refs/heads/united","pushedAt":"2023-03-18T06:32:27.000Z","pushType":"pr_merge","commitsCount":26,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #26 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #26 from henk717/united"}},{"before":"7f0fd92a590dcb9c3c8b45eb4aa93250b85dd8d7","after":"835cf50ebc381d37367336d94bce532e1cb0b15e","ref":"refs/heads/united","pushedAt":"2023-03-09T06:58:44.065Z","pushType":"pr_merge","commitsCount":8,"pusher":{"login":"pi6am","name":"Llama","path":"/pi6am","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/34464159?s=80&v=4"},"commit":{"message":"Merge pull request #25 from henk717/united\n\nMerge united","shortMessageHtmlLink":"Merge pull request #25 from henk717/united"}}],"hasNextPage":false,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEio3HPQA","startCursor":null,"endCursor":null}},"title":"Activity ยท pi6am/KoboldAI"}