Julien Simon commited on
Commit
3fdf87c
·
1 Parent(s): 75e81c7
Files changed (3) hide show
  1. .pre-commit-config.yaml +77 -0
  2. .pylintrc +2 -0
  3. results.py +29 -6
.pre-commit-config.yaml ADDED
@@ -0,0 +1,77 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ repos:
2
+ - repo: https://github.com/pre-commit/pre-commit-hooks
3
+ rev: v4.6.0
4
+ hooks:
5
+ - id: trailing-whitespace
6
+ - id: end-of-file-fixer
7
+ - id: check-added-large-files
8
+
9
+ - repo: https://github.com/psf/black
10
+ rev: 24.8.0
11
+ hooks:
12
+ - id: black
13
+ entry: bash -c 'black "$@"; git add -u' --
14
+
15
+ - repo: https://github.com/pycqa/isort
16
+ rev: 5.13.2
17
+ hooks:
18
+ - id: isort
19
+ args: ["--profile", "black"]
20
+ entry: bash -c 'isort "$@"; git add -u' --
21
+
22
+ - repo: https://github.com/PyCQA/flake8
23
+ rev: 7.1.1
24
+ hooks:
25
+ - id: flake8
26
+ args: [--max-line-length=120, "--extend-ignore=E501,E402"]
27
+
28
+ - repo: https://github.com/kynan/nbstripout
29
+ rev: 0.7.1
30
+ hooks:
31
+ - id: nbstripout
32
+
33
+ - repo: https://github.com/nbQA-dev/nbQA
34
+ rev: 1.8.7
35
+ hooks:
36
+ - id: nbqa-black
37
+ additional_dependencies: [black==23.3.0]
38
+ entry: bash -c 'nbqa-black "$@"; git add -u' --
39
+ - id: nbqa-isort
40
+ additional_dependencies: [isort==5.12.0]
41
+ entry: bash -c 'nbqa-isort "$@"; git add -u' --
42
+ - id: nbqa-flake8
43
+ additional_dependencies: [flake8==6.0.0]
44
+ args: [--max-line-length=120, "--extend-ignore=E501,E402,E231"]
45
+
46
+ - repo: https://github.com/mwouts/jupytext
47
+ rev: v1.16.4b
48
+ hooks:
49
+ - id: jupytext
50
+ args: [--sync, --pipe, black]
51
+ additional_dependencies:
52
+ - black==23.3.0
53
+
54
+ - repo: https://github.com/aws-cloudformation/cfn-lint
55
+ rev: v1.11.0 # Use the latest version available
56
+ hooks:
57
+ - id: cfn-lint
58
+ files: cloudformation/.*\.(json|yml|yaml)$
59
+
60
+ - repo: https://github.com/asottile/pyupgrade
61
+ rev: v3.17.0
62
+ hooks:
63
+ - id: pyupgrade
64
+ args: [--py310-plus]
65
+ entry: bash -c 'pyupgrade "$@"; git add -u' --
66
+
67
+ - repo: https://github.com/pre-commit/mirrors-mypy
68
+ rev: v1.11.2
69
+ hooks:
70
+ - id: mypy
71
+ entry: bash -c 'mypy "$@"; git add -u' --
72
+
73
+ - repo: https://github.com/PyCQA/pylint
74
+ rev: v3.2.6
75
+ hooks:
76
+ - id: pylint
77
+ args: [--rcfile=.pylintrc]
.pylintrc ADDED
@@ -0,0 +1,2 @@
 
 
 
1
+ [MESSAGES CONTROL]
2
+ disable=C0301,E0401,R0914
results.py CHANGED
@@ -1,3 +1,5 @@
 
 
1
  results = {
2
  "models": [
3
  {
@@ -27,7 +29,7 @@ results = {
27
  "status": "OK",
28
  "tokensPerSecond": "38",
29
  "notes": "",
30
- }
31
  ],
32
  },
33
  {
@@ -57,7 +59,7 @@ results = {
57
  "status": "OK",
58
  "tokensPerSecond": "38",
59
  "notes": "",
60
- }
61
  ],
62
  },
63
  {
@@ -334,7 +336,7 @@ results = {
334
  "quantization": "none",
335
  "container": "SGLang 0.2.13",
336
  "status": "OK",
337
- "tokensPerSecond": "45"
338
  },
339
  {
340
  "region": "AWS",
@@ -345,7 +347,7 @@ results = {
345
  "quantization": "none",
346
  "container": "vLLM 0.5.5",
347
  "status": "OK",
348
- "tokensPerSecond": "43.4"
349
  },
350
  {
351
  "region": "AWS",
@@ -476,8 +478,7 @@ results = {
476
  },
477
  ],
478
  },
479
- {"name": "Arcee-Spark",
480
- "modelType": "Qwen2 7B"},
481
  {
482
  "name": "Arcee-Lite",
483
  "modelType": "Qwen2 1.5B distilled from phi-3-medium 14B",
@@ -577,6 +578,28 @@ results = {
577
  "status": "OK",
578
  "tokensPerSecond": "160",
579
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
580
  ],
581
  },
582
  {
 
1
+ """Module containing model configuration results for various AI models and hardware setups."""
2
+
3
  results = {
4
  "models": [
5
  {
 
29
  "status": "OK",
30
  "tokensPerSecond": "38",
31
  "notes": "",
32
+ },
33
  ],
34
  },
35
  {
 
59
  "status": "OK",
60
  "tokensPerSecond": "38",
61
  "notes": "",
62
+ },
63
  ],
64
  },
65
  {
 
336
  "quantization": "none",
337
  "container": "SGLang 0.2.13",
338
  "status": "OK",
339
+ "tokensPerSecond": "45",
340
  },
341
  {
342
  "region": "AWS",
 
347
  "quantization": "none",
348
  "container": "vLLM 0.5.5",
349
  "status": "OK",
350
+ "tokensPerSecond": "43.4",
351
  },
352
  {
353
  "region": "AWS",
 
478
  },
479
  ],
480
  },
481
+ {"name": "Arcee-Spark", "modelType": "Qwen2 7B"},
 
482
  {
483
  "name": "Arcee-Lite",
484
  "modelType": "Qwen2 1.5B distilled from phi-3-medium 14B",
 
578
  "status": "OK",
579
  "tokensPerSecond": "160",
580
  },
581
+ {
582
+ "region": "us-west-2",
583
+ "instanceType": "g6e.2xlarge",
584
+ "cloud": "AWS",
585
+ "gpu": "1xNVIDIA L40S",
586
+ "gpuRAM": "48 GB",
587
+ "quantization": "none",
588
+ "container": "vLLM 0.5.5",
589
+ "status": "OK",
590
+ "tokensPerSecond": "146",
591
+ },
592
+ {
593
+ "region": "us-west-2",
594
+ "instanceType": "g6e.2xlarge",
595
+ "cloud": "AWS",
596
+ "gpu": "1xNVIDIA L40S",
597
+ "gpuRAM": "48 GB",
598
+ "quantization": "none",
599
+ "container": "SGLang 0.2.13",
600
+ "status": "OK",
601
+ "tokensPerSecond": "167",
602
+ },
603
  ],
604
  },
605
  {