Skip to content

Commit

Permalink
fix: groq max_tokens (#29)
Browse files Browse the repository at this point in the history
* fix: groq max_tokens

* test: use sequential execution for e2e cli tests
  • Loading branch information
gmickel authored Jul 28, 2024
1 parent 5d82bad commit 284615c
Show file tree
Hide file tree
Showing 3 changed files with 3 additions and 3 deletions.
2 changes: 1 addition & 1 deletion src/ai/model-config.ts
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ export const MODEL_CONFIGS: ModelSpecs = {
},
'llama-3.1-70b-versatile': {
contextWindow: 131072,
maxOutput: 8192,
maxOutput: 8000,
modelName: 'Llama 3.1 70B Groq',
pricing: { inputCost: 0.15, outputCost: 0.6 },
modelFamily: 'groq',
Expand Down
2 changes: 1 addition & 1 deletion tests/e2e/cli-commands.test.ts
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@ import { normalizePath } from '../../src/utils/normalize-path';

const execAsync = promisify(exec);

describe('CLI Commands', () => {
describe.sequential('CLI Commands', () => {
const cliPath = path.resolve(__dirname, '..', '..', 'src', 'cli', 'index.ts');
const testProjectPath = path.resolve(
__dirname,
Expand Down
2 changes: 1 addition & 1 deletion tests/unit/token-management.test.ts
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,7 @@ describe('Token Management', () => {
const specs = getModelConfig('llama-3.1-70b-versatile');
expect(specs).toEqual({
contextWindow: 131072,
maxOutput: 8192,
maxOutput: 8000,
modelName: 'Llama 3.1 70B Groq',
pricing: { inputCost: 0.15, outputCost: 0.6 },
modelFamily: 'groq',
Expand Down

0 comments on commit 284615c

Please sign in to comment.