From 94cbdc664459704304c5a612ea5aa5c7a40a0117 Mon Sep 17 00:00:00 2001 From: Martin Yuan Date: Tue, 3 Oct 2023 12:16:48 -0700 Subject: [PATCH] Add Meta llama as a submodule. (#560) Summary: Pull Request resolved: https://github.com/pytorch/executorch/pull/560 imported-using-ghimport Test Plan: Imported from OSS Reviewed By: cccclai Differential Revision: D49734020 Pulled By: iseeyuan fbshipit-source-id: b6eea7d3da570bb7c8e5f7ea09916e414cc204f3 --- .gitmodules | 3 +++ examples/third-party/llama | 1 + 2 files changed, 4 insertions(+) create mode 160000 examples/third-party/llama diff --git a/.gitmodules b/.gitmodules index 826924b71cd..4b783f4abc0 100644 --- a/.gitmodules +++ b/.gitmodules @@ -34,3 +34,6 @@ [submodule "third-party/flatcc"] path = third-party/flatcc url = https://github.com/dvidelabs/flatcc.git +[submodule "examples/third-party/llama"] + path = examples/third-party/llama + url = https://github.com/facebookresearch/llama.git diff --git a/examples/third-party/llama b/examples/third-party/llama new file mode 160000 index 00000000000..b00a461a658 --- /dev/null +++ b/examples/third-party/llama @@ -0,0 +1 @@ +Subproject commit b00a461a6582196d8f488c73465f6c87f384a052