diff --git a/.github/workflows/test-action.yml b/.github/workflows/test-action.yml index 5a1f111..63e44ca 100644 --- a/.github/workflows/test-action.yml +++ b/.github/workflows/test-action.yml @@ -46,7 +46,7 @@ jobs: } >> $GITHUB_OUTPUT - uses: ./ name: "Code Review" - id: review + id: review1 with: apiKey: ${{ secrets.API_KEY }} githubToken: ${{ secrets.GITHUB_TOKEN }} @@ -62,3 +62,39 @@ jobs: ${{ steps.get_diff.outputs.pull_request_diff }} pullRequestChunkSize: "3500" logLevel: "DEBUG" + - uses: ./ + name: "Code Review" + id: review2 + with: + apiKey: ${{ secrets.API_KEY }} + githubToken: ${{ secrets.GITHUB_TOKEN }} + githubRepository: ${{ github.repository }} + githubPullRequestNumber: ${{ github.event.pull_request.number }} + gitCommitHash: ${{ github.event.pull_request.head.sha }} + repoId: "meta-llama/Llama-2-13b-chat-hf" + temperature: "0.2" + maxNewTokens: "250" + topK: "50" + topP: "0.95" + pullRequestDiff: |- + ${{ steps.get_diff.outputs.pull_request_diff }} + pullRequestChunkSize: "3500" + logLevel: "DEBUG" + - uses: ./ + name: "Code Review" + id: review3 + with: + apiKey: ${{ secrets.API_KEY }} + githubToken: ${{ secrets.GITHUB_TOKEN }} + githubRepository: ${{ github.repository }} + githubPullRequestNumber: ${{ github.event.pull_request.number }} + gitCommitHash: ${{ github.event.pull_request.head.sha }} + repoId: "codellama/CodeLlama-13b-hf" + temperature: "0.2" + maxNewTokens: "250" + topK: "50" + topP: "0.95" + pullRequestDiff: |- + ${{ steps.get_diff.outputs.pull_request_diff }} + pullRequestChunkSize: "3500" + logLevel: "DEBUG" diff --git a/README.md b/README.md index eadc4eb..3d7fa02 100644 --- a/README.md +++ b/README.md @@ -69,7 +69,7 @@ jobs: githubRepository: ${{ github.repository }} githubPullRequestNumber: ${{ github.event.pull_request.number }} gitCommitHash: ${{ github.event.pull_request.head.sha }} - repoId: "meta-llama/Llama-2-7b-chat-hf" + repoId: "codellama/CodeLlama-7b-hf" temperature: "0.2" maxNewTokens: "250" topK: "50" diff --git a/action.yml b/action.yml index d36b927..9aa39a2 100644 --- a/action.yml +++ b/action.yml @@ -23,7 +23,7 @@ inputs: repoId: description: "LLM model" required: true - default: "meta-llama/Llama-2-7b-chat-hf" + default: "codellama/CodeLlama-7b-hf" maxNewTokens: description: "The amount of new tokens to be generated, this does not include the input length it is a estimate of the size of generated text you want. Each new tokens slows down the request, so look for balance between response times and length of text generated." required: false