{"payload":{"feedbackUrl":"https://github.com/orgs/community/discussions/53140","repo":{"id":689773665,"defaultBranch":"main","name":"llamafile","ownerLogin":"Mozilla-Ocho","currentUserCanPush":false,"isFork":false,"isEmpty":false,"createdAt":"2023-09-10T21:12:32.000Z","ownerAvatar":"https://avatars.githubusercontent.com/u/117940224?v=4","public":true,"private":false,"isOrgOwned":true},"refInfo":{"name":"","listCacheKey":"v0:1716647070.0","currentOid":""},"activityList":{"items":[{"before":"7d8dd1b33fd54e9e54d4ad8074f8df64e547b75d","after":"9cd8d70942a049ba3c3bddd12e87e1fb599fbd49","ref":"refs/heads/main","pushedAt":"2024-06-01T15:51:45.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Update sever README build/testing instructions (#461)","shortMessageHtmlLink":"Update sever README build/testing instructions (#461)"}},{"before":"293a5284c49318bb2cef4ab781331edce3f2243c","after":"7d8dd1b33fd54e9e54d4ad8074f8df64e547b75d","ref":"refs/heads/main","pushedAt":"2024-06-01T10:09:40.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Upgrade to Cosmopolitan v3.3.10 (#460)\n\nNeeded to fix https://github.com/Mozilla-Ocho/llamafile/issues/446 on windows","shortMessageHtmlLink":"Upgrade to Cosmopolitan v3.3.10 (#460)"}},{"before":"73088c3bb0e3143fec0d356feb97a0cacd2c0d70","after":"293a5284c49318bb2cef4ab781331edce3f2243c","ref":"refs/heads/main","pushedAt":"2024-05-30T00:10:31.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Performance improvements on Arm for legacy and k-quants (#453)","shortMessageHtmlLink":"Performance improvements on Arm for legacy and k-quants (#453)"}},{"before":"31419d0b718f318ab23ab40eeb10a170e0eb2edc","after":"73088c3bb0e3143fec0d356feb97a0cacd2c0d70","ref":"refs/heads/main","pushedAt":"2024-05-29T17:38:11.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"github: delete question in favor of link to discussion [no ci] (#457)","shortMessageHtmlLink":"github: delete question in favor of link to discussion [no ci] (#457)"}},{"before":"397175e673c4334962f446d9470e3bceefc88fb0","after":"31419d0b718f318ab23ab40eeb10a170e0eb2edc","ref":"refs/heads/main","pushedAt":"2024-05-29T07:24:34.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"github: add ci (#454)","shortMessageHtmlLink":"github: add ci (#454)"}},{"before":"92be52a3bbde8366becff2cdd550cc6a249f7c43","after":"397175e673c4334962f446d9470e3bceefc88fb0","ref":"refs/heads/main","pushedAt":"2024-05-26T11:55:21.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"github: add mention of strace and ftrace (#449)","shortMessageHtmlLink":"github: add mention of strace and ftrace (#449)"}},{"before":"ba7193043ba5c51fde6a5e146883dc87aaf07a85","after":"92be52a3bbde8366becff2cdd550cc6a249f7c43","ref":"refs/heads/main","pushedAt":"2024-05-26T11:44:11.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"actions: add labeler + editorconfig github actions (#443)\n\n* actions: add labler + editorconfig github actions\r\n\r\n* Update labeler.yml","shortMessageHtmlLink":"actions: add labeler + editorconfig github actions (#443)"}},{"before":"076dfb0dae2169abb62f490218b5053f37f61cfc","after":"ba7193043ba5c51fde6a5e146883dc87aaf07a85","ref":"refs/heads/main","pushedAt":"2024-05-26T10:56:49.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"github: delete assignees and about --> description (#448)","shortMessageHtmlLink":"github: delete assignees and about --> description (#448)"}},{"before":"81cfbcf48ee037912eed78e34cc214dac0d2a6d5","after":"076dfb0dae2169abb62f490218b5053f37f61cfc","ref":"refs/heads/main","pushedAt":"2024-05-26T01:16:46.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"github: add issue templates (#442)","shortMessageHtmlLink":"github: add issue templates (#442)"}},{"before":"ea2a96e5bf8216d002ff40d3283cce4f2100b181","after":"81cfbcf48ee037912eed78e34cc214dac0d2a6d5","ref":"refs/heads/main","pushedAt":"2024-05-25T14:24:24.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Release llamafile v0.8.6","shortMessageHtmlLink":"Release llamafile v0.8.6"}},{"before":"b79ecf465befa8018e3331720372917454097a90","after":"ea2a96e5bf8216d002ff40d3283cce4f2100b181","ref":"refs/heads/main","pushedAt":"2024-05-25T09:17:40.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Disable GPU in llava-quantize","shortMessageHtmlLink":"Disable GPU in llava-quantize"}},{"before":"e67571914779c847233c2ea1e05c587769298f7f","after":"b79ecf465befa8018e3331720372917454097a90","ref":"refs/heads/main","pushedAt":"2024-05-25T08:45:56.000Z","pushType":"push","commitsCount":4,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Release llamafile v0.8.5","shortMessageHtmlLink":"Release llamafile v0.8.5"}},{"before":"4451c6d98f31325c9eae3e4be0351883096a831d","after":"e67571914779c847233c2ea1e05c587769298f7f","ref":"refs/heads/main","pushedAt":"2024-05-24T20:06:18.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Make some more benchmark tool fixes","shortMessageHtmlLink":"Make some more benchmark tool fixes"}},{"before":"91dd4d371ef383a0c22b7c94aea963863ba8c30d","after":"4451c6d98f31325c9eae3e4be0351883096a831d","ref":"refs/heads/main","pushedAt":"2024-05-24T16:38:32.000Z","pushType":"push","commitsCount":4,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Reclaim mapped memory","shortMessageHtmlLink":"Reclaim mapped memory"}},{"before":"9206719486eadedd850c06c4ec3b820742d12de8","after":"91dd4d371ef383a0c22b7c94aea963863ba8c30d","ref":"refs/heads/main","pushedAt":"2024-05-23T13:42:49.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Add missing CPUID checks","shortMessageHtmlLink":"Add missing CPUID checks"}},{"before":"7cb15c64e14813bc65fa1aecd53262b6a05dd2c5","after":"9206719486eadedd850c06c4ec3b820742d12de8","ref":"refs/heads/main","pushedAt":"2024-05-23T13:07:53.000Z","pushType":"push","commitsCount":3,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Add llama-bench command (cpu mode only)\n\nFor example:\n\n #!/bin/sh\n cd ~/llamafile\n make -j32 o//llama.cpp/llama-bench/llama-bench || exit\n o//llama.cpp/llama-bench/llama-bench \\\n $(for f in $(ls -S /weights/TinyLlama-1.1B-Chat-v1.0.*.gguf \\\n /weights/mixtral-8x7b-instruct-v0.1.*.gguf); do\n echo -m $f\n done) \\\n \"$@\"\n\nI store that to ~/bin/bench-llamafile.sh","shortMessageHtmlLink":"Add llama-bench command (cpu mode only)"}},{"before":"d228e01d70a7b91bf04dbf63428646f3f173b888","after":"7cb15c64e14813bc65fa1aecd53262b6a05dd2c5","ref":"refs/heads/main","pushedAt":"2024-05-23T13:05:41.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Another performance optimization for Zen4 + refactoring (#435)\n\n* Matrix multiplications optimized for Zen4 + simplify\r\n\r\n* A minor optimization\r\n\r\n* Parallelize quantization also in ggml_compute_forward_mul_mat_id\r\n\r\nI noticed that my AVX2 implemetation of Q8_K quantization\r\n(needed by k- and i-quants) has been lost. jart has counteracted\r\nthis by parallelizing quantization, but only in\r\nggml_compute_forward_mul_mat. Adding the exact same technique to\r\nggml_compute_forward_mul_mat_id results in a 5-6% performance\r\nimprovement for Mixtral8x7B. This is on top of the improvement due\r\nto the better matrix multiplication implementation.","shortMessageHtmlLink":"Another performance optimization for Zen4 + refactoring (#435)"}},{"before":"b3aa97d512648f2ca262f38b67a6b5e0adbee30d","after":"d228e01d70a7b91bf04dbf63428646f3f173b888","ref":"refs/heads/main","pushedAt":"2024-05-22T06:26:58.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Sync with llama.cpp upstream\n\nYou can now use Phi-3-Medium-128k although not with CUDA yet, since\nggml-cuda is broken at HEAD. This syncs one commit back for cuda so\nCPU and metal should be fine. Brain16 prediction now goes faster on\nCPU too, thanks to fixing a minor mistake in the tinyBLAS config.","shortMessageHtmlLink":"Sync with llama.cpp upstream"}},{"before":"87d4ce1a90b7f31fc6ce69eee7186321c432ade9","after":"b3aa97d512648f2ca262f38b67a6b5e0adbee30d","ref":"refs/heads/main","pushedAt":"2024-05-21T17:52:03.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Fix typo in llama.h (#354)\n\nindicies -> indices","shortMessageHtmlLink":"Fix typo in llama.h (#354)"}},{"before":"938cf72b4f562e3dac219a1d38d98202310aa07d","after":"87d4ce1a90b7f31fc6ce69eee7186321c432ade9","ref":"refs/heads/main","pushedAt":"2024-05-21T17:41:12.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Fix f16 cpuid check\n\nFixes #413","shortMessageHtmlLink":"Fix f16 cpuid check"}},{"before":"00e4f72b85b824a8dea98803b855dea2b50c02da","after":"938cf72b4f562e3dac219a1d38d98202310aa07d","ref":"refs/heads/main","pushedAt":"2024-05-21T05:26:48.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Faster AVX2 matrix multiplications for MoE models (#428)","shortMessageHtmlLink":"Faster AVX2 matrix multiplications for MoE models (#428)"}},{"before":"64fbffcca2af41236e5ac29bae74deab088df2e7","after":"00e4f72b85b824a8dea98803b855dea2b50c02da","ref":"refs/heads/main","pushedAt":"2024-05-21T01:36:50.000Z","pushType":"push","commitsCount":2,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Enable GGML_CUDA_FORCE_MMQ in tinyBLAS mode\n\nFixes #403","shortMessageHtmlLink":"Enable GGML_CUDA_FORCE_MMQ in tinyBLAS mode"}},{"before":"49cc13ca2bb381164abcfca2a86782062fda2c7a","after":"64fbffcca2af41236e5ac29bae74deab088df2e7","ref":"refs/heads/main","pushedAt":"2024-05-20T19:39:15.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Update llama.cpp (#427)","shortMessageHtmlLink":"Update llama.cpp (#427)"}},{"before":"3a0cf4e1b43f1d082aa206fb056bcffd0367630c","after":"49cc13ca2bb381164abcfca2a86782062fda2c7a","ref":"refs/heads/main","pushedAt":"2024-05-17T09:43:58.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Updated README with instructions to load models from third-party apps (#417)","shortMessageHtmlLink":"Updated README with instructions to load models from third-party apps ("}},{"before":"d5f614c9d7d1efdf6d40a8812d7f148f41aa1072","after":"3a0cf4e1b43f1d082aa206fb056bcffd0367630c","ref":"refs/heads/main","pushedAt":"2024-05-16T23:54:16.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Fix typo in README.md (#407)\n\nSee also #376","shortMessageHtmlLink":"Fix typo in README.md (#407)"}},{"before":"c660d38da613969a7c9be0d69cc56a10f878209b","after":null,"ref":"refs/heads/add-embedding-models","pushedAt":"2024-05-16T14:55:44.000Z","pushType":"branch_deletion","commitsCount":0,"pusher":{"login":"k8si","name":"Kate Silverstein","path":"/k8si","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3207674?s=80&v=4"}},{"before":"966dd23477f934ad41f2e7840eb1f3e0d7fa7caa","after":"d5f614c9d7d1efdf6d40a8812d7f148f41aa1072","ref":"refs/heads/main","pushedAt":"2024-05-16T14:55:39.000Z","pushType":"pr_merge","commitsCount":2,"pusher":{"login":"k8si","name":"Kate Silverstein","path":"/k8si","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3207674?s=80&v=4"},"commit":{"message":"Merge pull request #422 from Mozilla-Ocho/add-embedding-models\n\nadd text embedding models to 'other example llamafiles' table","shortMessageHtmlLink":"Merge pull request #422 from Mozilla-Ocho/add-embedding-models"}},{"before":null,"after":"c660d38da613969a7c9be0d69cc56a10f878209b","ref":"refs/heads/add-embedding-models","pushedAt":"2024-05-16T14:39:16.000Z","pushType":"branch_creation","commitsCount":0,"pusher":{"login":"k8si","name":"Kate Silverstein","path":"/k8si","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/3207674?s=80&v=4"},"commit":{"message":"add text embedding models to 'other example llamafiles' table","shortMessageHtmlLink":"add text embedding models to 'other example llamafiles' table"}},{"before":"a86e7ce6bf19a976957429945389add858ce60eb","after":"966dd23477f934ad41f2e7840eb1f3e0d7fa7caa","ref":"refs/heads/main","pushedAt":"2024-05-14T05:27:15.000Z","pushType":"push","commitsCount":1,"pusher":{"login":"stlhood","name":"Stephen Hood","path":"/stlhood","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/42821?s=80&v=4"},"commit":{"message":"Update HF links to point to new Mozilla account","shortMessageHtmlLink":"Update HF links to point to new Mozilla account"}},{"before":"d4099feff737f3be9a4bed017f1315a4d2cb773a","after":"a86e7ce6bf19a976957429945389add858ce60eb","ref":"refs/heads/main","pushedAt":"2024-05-13T04:33:35.000Z","pushType":"pr_merge","commitsCount":1,"pusher":{"login":"jart","name":"Justine Tunney","path":"/jart","primaryAvatarUrl":"https://avatars.githubusercontent.com/u/49262?s=80&v=4"},"commit":{"message":"Add Script To Upgrade llamafile Archives (#412)","shortMessageHtmlLink":"Add Script To Upgrade llamafile Archives (#412)"}}],"hasNextPage":true,"hasPreviousPage":false,"activityType":"all","actor":null,"timePeriod":"all","sort":"DESC","perPage":30,"cursor":"djE6ks8AAAAEWcqc7AA","startCursor":null,"endCursor":null}},"title":"Activity · Mozilla-Ocho/llamafile"}