Code Monkey home page Code Monkey logo

yolov8_aimbot's Introduction

Sunone Aimbot

Python Version License MIT Github stars Discord server

Overview

Sunone Aimbot is an AI-powered aim bot for first-person shooter games. It leverages the YOLOv8 and YOLOv10 models, PyTorch, and various other tools to automatically target and aim at enemies within the game. The AI model in repository has been trained on more than 30,000 images from popular first-person shooter games like Warface, Destiny 2, Battlefield (all series), Fortnite, The Finals, CS2 and more.

Warning

Use it at your own risk, we do not guarantee that you may be blocked!

Note

This application only works on Nvidia graphics cards. AMD support is testing. See AI_enable_AMD option. The recommended graphics card for starting and more productive and stable operation starts with the rtx 20 series.

Requirements

Before you get started, make sure you have the following prerequisites installed and pay attention to the versions in Tested Environment block, this may cause errors in launching the aimbot.

Tested Environment

The Sunone Aimbot has been tested on the following environment:

Windows10 and 11(priority)
Python:3.11.6
CUDA:12.4
TensorRT:10.0.1
Ultralytics:8.2.49
GitHub AI Model:0.4.1 (YOLOv8)
Boosty AI Model:0.5.8 (YOLOv10)

Options

The behavior of the aimbot can be configured via the config.ini file. Here are the available options:

Object Search window resolution:

  • detection_window_width int: Horizontal resolution of the object search window.
  • detection_window_height int: Vertical resolution of the object search window.

Bettercam capture method:

  • Bettercam_capture bool: Use Bettercam to capture images from the screen.
  • bettercam_capture_fps int: Specific fps value for screen capture.
  • bettercam_monitor_id int: Id of the monitor from which the images will be captured.
  • bettercam_gpu_id int: Id of the GPU to be used for image capture

Obs capture method:

  • Obs_capture bool: Use Obs to capture images from the screen.
  • Obs_camera_id str or int: auto or number of Virtual Camera ID.
  • Obs_capture_fps int: Specific fps value for screen capture.

Aim:

  • body_y_offset float: Allows correction of y coordinates inside the body detected box if head is not detected.
  • hideout_targets bool: Allows shooting at targets on the range (for example in warface on the polygon or in aimlabs).
  • disable_headshot bool: Disable head targerting.
  • disable_prediction bool: Disable target position prediction.
  • third_person bool: Turn on the third-person game mode. (sunxds_0.5.7+)

Hot keys:

  • The names of all the keys are here. Type None is empty button.
  • hotkey_targeting str: Aiming at the target. Supports multi keys, for example hotkey_targeting = RightMouseButton,X2MouseButton
  • hotkey_exit str: Exit.
  • hotkey_pause str: Pause AIM.
  • hotkey_reload_config str: Reload config.

Mouse:

  • mouse_dpi int: Mouse DPI.
  • mouse_sensitivity float: Aim sensitivity.
  • mouse_fov_width int: The current horizontal value of the viewing angle in the game.
  • mouse_fov_height int: The current vertical value of the viewing angle in the game.
  • mouse_lock_target bool: True: Press once to permanently aim at the target, press again to turn off the aiming. False: Hold down the button to constantly aim at the target.
  • mouse_auto_aim bool: Automatic targeting.
  • mouse_ghub bool: Uses Logitech GHUB exploit for mouse movement. If the value is False, native win32 library is used for movement.

Shooting:

  • auto_shoot bool: Automatic shooting. (For some games need arduino).
  • triggerbot bool: Automatic shooting at a target if it is in the scope, requires the mouse_auto_shoot option enabled, and aiming will also be automatically turned off.
  • force_click bool: Shooting will be performed even if the sight is not located within the object.
  • bScope_multiplier float: The multiplier of the target trigger size.

Arduino:

  • arduino_move bool: Sends a command to the arduino to move the mouse.
  • arduino_shoot bool: Sends a command to the arduino to fire with the mouse.
  • arduino_port str: Arduino COM port. Use COM1 or COM2 ... or auto.
  • arduino_baudrate int: Custom Arduino baudrate.
  • arduino_16_bit_mouse bool: Send 16 bit data to the arduino port to move the mouse.

AI:

  • AI_model_name str: AI model name.
  • AI_model_image_size int: AI model image size.
  • AI_conf float: How many percent is AI sure that this is the right goal.
  • AI_device int or str: Device to run on, 0, 1... or cpu.
  • AI_enable_AMD bool: Enable support Amd GPUs. Install ROCm, Zluda and PyTorch. See AMD docs.
  • AI_mouse_net bool: Use a neural network to calculate mouse movements. See this repository.

Overlay:

  • show_overlay bool: Enables the overlay. It is not recommended for gameplay, only for debugging.
  • overlay_show_borders bool: Displaying the borders of the overlay.
  • overlay_show_boxes bool: Display of boxes.
  • overlay_show_target_line bool: Displaying the line to the target.
  • overlay_show_target_prediction_line bool: Displaying the predictive line to the target.
  • overlay_show_labels bool: Displaying label names.
  • overlay_show_conf bool: Displaying the label names as well as the confidence level.

Debug window:

  • show_window bool: Shows the OpenCV2 window for visual feedback.
  • show_detection_speed bool: Displays speed information inside the debug window.
  • show_window_fps bool: Displays FPS in the corner.
  • show_boxes bool: Displays detectable objects.
  • show_labels bool: Displays the name of the detected object.
  • show_conf bool: Displays object confidence threshold for detection.
  • show_target_line bool: Shows the mouse finishing line.
  • show_target_prediction_line bool: Show mouse prediction line.
  • show_bScope_box bool: Show the trigger box for auto shooting.
  • show_history_points bool: Show history points.
  • debug_window_always_on_top bool: The debug window will always be on top of other windows.
  • spawn_window_pos_x int: When the debugging window starts, it takes the x position.
  • spawn_window_pos_y int: When the debugging window starts, it takes the y position.
  • debug_window_scale_percent int: Adjusts the size of the debug window.
  • The names of the debugging window can be written in the file window_names.txt they will be randomly selected.

AI Models

  • *.pt: Default AI model.
  • *.onnx: The model is optimized to run on processors.
  • *.engine: Final exported model, which is faster than the previous two.

Export .pt model to .engine

  1. All commands are executed in the console window:
  2. First, go to the aimbot directory using the command:
cd C:\Users\your_username\downloads\sunone_aimbot-main
  1. Then export the model from the .pt format in .engine format.
yolo export model="models/sunxds_0.4.1.pt" format=engine device=0 imgsz=480 half=True
  • model="model_path/model_name.pt": Path to model.
  • format=engine: TensorRT model format.
  • half=true: Use Half-precision floating-point format.
  • device=0: GPU id.
  • workspace=8: GPU max video memory.
  • verbose=False: Debug stuff. Convenient function, can show errors when exporting.
  1. Each model has its own image size with which it was trained, export only with the image size with which it was published.

Notes / Recommendations

  • Limit the maximum value of frames per second in the game in which you will use it. And also do not set the screen resolution to high. Do not overload the graphics card.
  • Do not set high graphics settings in games.
  • Limit the browser (try not to watch YouTube while playing and working AI at the same time, for example (of course if you don't have a super duper graphics card)) and so on, which loads the video card.
  • Try to use TensorRT for acceleration. .pt model is good, but does not have as much speed as .engine.
  • Turn off the cv2 debug window, this saves system resources.
  • Do not increase the object search window resolution, this may affect your search speed.
  • If you have started the application and nothing happens, it may be working, close it with the F2 key and change the show_window option to True in the file config.ini to make sure that the application is working.

Support the project

I will post new models here.

License

This project is licensed under the MIT License. See LICENSE for details

yolov8_aimbot's People

Contributors

kai0kenx20 avatar sunoner avatar trendesetter avatar weser87 avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

yolov8_aimbot's Issues

Mouse doesnt move with native mouse set to false

Hi,

For some reason when not using native mouse mode, the mouse does not move for me at all. Like it is unable to target anything with right click, even with detections available. Any idea what could be causing this? I use a logitech mouse myself with ghub and perhaps it conflicts with the logitech dll you seem to be using, I havent been able to get it to work at all though. Thank you for your work, I did also send you a message on discord regarding this.

the updated version does not start

Hi! I downloaded the code that you updated yesterday, but it doesn't start and doesn't give any errors, just nothing happens, dxcam doesn't start. although your old version is working properly!

I use names that will soon no longer be available.

On startup I get this error, in a constant loop. The entire log is filled with this error. I think that you should pay attention to this and change the names of the variables before the library is released.
WARNING ⚠️ 'boxes' is deprecated and will be removed in 'ultralytics 8.2' in the future. Please use 'show_boxes' instead.

Ошибка файлов при запуске

image
почему-то выдаёт ошибку при самом последнем вводе строки
yolo export model="models/sunxds_0.2.1_nano_320.pt" format=engine half=true device=0 imgsz=320

TypeError: 'NoneType' object does not support the context manager protocol

C:\Users\User\Desktop\yolov8_aimbot-main>yolo export model="models/sunxds_0.2.1_nano_320.pt" format=engine half=true device=0 imgsz=320
Ultralytics YOLOv8.0.225 🚀 Python-3.11.6 torch-2.1.1+cu121 CUDA:0 (NVIDIA GeForce GTX 750 Ti, 4096MiB)
Model summary (fused): 168 layers, 3007598 parameters, 0 gradients, 8.1 GFLOPs

PyTorch: starting from 'models\sunxds_0.2.1_nano_320.pt' with input shape (1, 3, 320, 320) BCHW and output shape(s) (1, 14, 2100) (5.9 MB)
requirements: Ultralytics requirement ['onnx>=1.12.0'] not found, attempting AutoUpdate...
Collecting onnx>=1.12.0
Downloading onnx-1.15.0-cp311-cp311-win_amd64.whl.metadata (15 kB)
Requirement already satisfied: numpy in c:\users\user\appdata\local\programs\python\python311\lib\site-packages (from onnx>=1.12.0) (1.26.2)
Requirement already satisfied: protobuf>=3.20.2 in c:\users\user\appdata\local\programs\python\python311\lib\site-packages (from onnx>=1.12.0) (4.25.1)
Downloading onnx-1.15.0-cp311-cp311-win_amd64.whl (14.3 MB)
---------------------------------------- 14.3/14.3 MB 11.9 MB/s eta 0:00:00
Installing collected packages: onnx
Successfully installed onnx-1.15.0

requirements: AutoUpdate success ✅ 10.4s, installed 1 package: ['onnx>=1.12.0']
requirements: ⚠️ Restart runtime or rerun command for updates to take effect

ONNX: starting export with onnx 1.15.0 opset 17...
ONNX: simplifying with onnxsim 0.4.35...
ONNX: export success ✅ 12.0s, saved as 'models\sunxds_0.2.1_nano_320.onnx' (5.8 MB)

TensorRT: starting export with TensorRT 8.6.1...
[12/10/2023-03:51:22] [TRT] [I] [MemUsageChange] Init CUDA: CPU +1, GPU +0, now: CPU 10223, GPU 815 (MiB)
[12/10/2023-03:51:29] [TRT] [I] [MemUsageChange] Init builder kernel library: CPU +130, GPU +0, now: CPU 11295, GPU 815 (MiB)
[12/10/2023-03:51:29] [TRT] [I] ----------------------------------------------------------------
[12/10/2023-03:51:29] [TRT] [I] Input filename: models\sunxds_0.2.1_nano_320.onnx
[12/10/2023-03:51:29] [TRT] [I] ONNX IR version: 0.0.8
[12/10/2023-03:51:29] [TRT] [I] Opset version: 17
[12/10/2023-03:51:29] [TRT] [I] Producer name: pytorch
[12/10/2023-03:51:29] [TRT] [I] Producer version: 2.1.1
[12/10/2023-03:51:29] [TRT] [I] Domain:
[12/10/2023-03:51:29] [TRT] [I] Model version: 0
[12/10/2023-03:51:29] [TRT] [I] Doc string:
[12/10/2023-03:51:29] [TRT] [I] ----------------------------------------------------------------
[12/10/2023-03:51:29] [TRT] [W] onnx2trt_utils.cpp:374: Your ONNX model has been generated with INT64 weights, while TensorRT does not natively support INT64. Attempting to cast down to INT32.
TensorRT: input "images" with shape(1, 3, 320, 320) DataType.HALF
TensorRT: output "output0" with shape(1, 14, 2100) DataType.HALF
TensorRT: building FP16 engine as models\sunxds_0.2.1_nano_320.engine
[12/10/2023-03:51:29] [TRT] [I] BuilderFlag::kTF32 is set but hardware does not support TF32. Disabling TF32.
[12/10/2023-03:51:29] [TRT] [I] Graph optimization time: 0.0508836 seconds.
[12/10/2023-03:51:29] [TRT] [I] BuilderFlag::kTF32 is set but hardware does not support TF32. Disabling TF32.
[12/10/2023-03:51:29] [TRT] [I] Local timing cache in use. Profiling results in this builder pass will not be stored.
[12/10/2023-03:51:29] [TRT] [E] 10: Could not find any implementation for node /model.0/conv/Conv.
[12/10/2023-03:51:29] [TRT] [E] 10: [optimizer.cpp::nvinfer1::builder::cgraph::LeafCNode::computeCosts::3869] Error Code 10: Internal Error (Could not find any implementation for node /model.0/conv/Conv.)
TensorRT: export failure ❌ 19.1s: 'NoneType' object does not support the context manager protocol
Traceback (most recent call last):
File "", line 198, in run_module_as_main
File "", line 88, in run_code
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Scripts\yolo.exe_main
.py", line 7, in
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\cfg_init
.py", line 448, in entrypoint
getattr(model, mode)(**overrides) # default args from model
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\model.py", line 310, in export
return Exporter(overrides=args, _callbacks=self.callbacks)(model=self.model)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 252, in call
f[1], _ = self.export_engine()
^^^^^^^^^^^^^^^^^^^^
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 122, in outer_func
raise e
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 117, in outer_func
f, model = inner_func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\User\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 631, in export_engine
with builder.build_engine(network, config) as engine, open(f, 'wb') as t:
TypeError: 'NoneType' object does not support the context manager protocol

yolo export format=engine,No response

yolo export model="models/sunxds_0.2.1_nano_320.pt" format=engine half=true device=0 imgsz=320
WARNING ⚠️ TensorRT requires GPU export, automatically assigning device=0
Ultralytics YOLOv8.0.222 🚀 Python-3.11.6 torch-2.1.1+cu121 CUDA:0 (NVIDIA GeForce RTX 2080 Ti, 11264MiB)
Model summary (fused): 168 layers, 3007598 parameters, 0 gradients, 8.1 GFLOPs

PyTorch: starting from 'models\sunxds_0.2.1_nano_320.pt' with input shape (1, 3, 320, 320) BCHW and output shape(s) (1, 14, 2100) (5.9 MB)

ONNX: starting export with onnx 1.15.0 opset 17...

And It's over.No response

по поводу amd

Подскажи пожалуйста, как запустить на amd, что скачать нужно, сабку если что оформлю на бусти, гуглил альтернативу на амд, показывал рком, но не воркает, он не видит cuda через кмд когда вписываю команду :(

( raise ValueError(f"Invalid CUDA 'device={device}' requested."
ValueError: Invalid CUDA 'device=0' requested. Use 'device=cpu' or pass valid CUDA device(s) if available, i.e. 'device=0' or 'device=0,1,2,3' for Multi-GPU.

torch.cuda.is_available(): False
torch.cuda.device_count(): 0
os.environ['CUDA_VISIBLE_DEVICES']: None
See https://pytorch.org/get-started/locally/ for up-to-date torch install instructions if no CUDA devices are seen by torch. )
Screenshot_25

sorry Why does it exit on its own?(cuda问题 已解决)

E:\yolo\yolov8_aimbot-main>python run.py
PT Model loaded.
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT
Loading models\sunxds_0.2.1_640.engine for TensorRT inference...
[12/12/2023-20:23:47] [TRT] [I] Loaded engine size: 24 MiB
[12/12/2023-20:23:47] [TRT] [W] Using an engine plan file across different models of devices is not recommended and is likely to affect performance or even cause errors.
[12/12/2023-20:23:47] [TRT] [I] [MemUsageChange] TensorRT-managed allocation in engine deserialization: CPU +0, GPU +21, now: CPU 0, GPU 21 (MiB)
[12/12/2023-20:23:47] [TRT] [I] [MemUsageChange] TensorRT-managed allocation in IExecutionContext creation: CPU +0, GPU +29, now: CPU 0, GPU 50 (MiB)

E:\yolo\yolov8_aimbot-main>

модель не найдена

исправил все и пишет теперь вот че C:\Users\igoty\Desktop\yolov8_aimbot-main>python main.py
Traceback (most recent call last):
File "C:\Users\igoty\Desktop\yolov8_aimbot-main\main.py", line 6, in
import win32con, win32api
ModuleNotFoundError: No module named 'win32con' че делать ?

проблема с ультрачетотам)

Microsoft Windows [Version 10.0.19045.3693]
(c) Корпорация Майкрософт (Microsoft Corporation). Все права защищены.

C:\Users\igoty\Desktop\yolov8_aimbot-main>checks.py
install ultralytics: pip install ultralytics

C:\Users\igoty\Desktop\yolov8_aimbot-main>

как установить скажите пожалуйста

Не экспортируется.

raise ValueError(f"Invalid export format='{fmt}'. Valid formats are {fmts}")
ValueError: Invalid export format='pt'. Valid formats are ('torchscript', 'onnx', 'openvino', 'engine', 'coreml', 'saved_model', 'pb', 'tflite', 'edgetpu', 'tfjs', 'paddle', 'ncnn')

WARNING ⚠️

WARNING ⚠️ 'boxes' is deprecated and will be removed in 'ultralytics 8.2' in the future. Please use 'show_boxes' instead ?

Error

How do I resolve this?

To change capture parameters you can manually delete the old object using del obj.
Traceback (most recent call last):
File "main.py", line 217, in
init()
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "main.py", line 92, in init
result = model(
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\ultralytics\engine\model.py", line 98, in call
return self.predict(source, stream, **kwargs)
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\torch\autograd\grad_mode.py", line 27, in decorate_context
return func(*args, **kwargs)
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\ultralytics\engine\model.py", line 237, in predict
self.predictor = predictor(overrides=overrides, callbacks=self.callbacks)
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\ultralytics\engine\predictor.py", line 86, in init
self.args = get_cfg(cfg, overrides)
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\ultralytics\cfg_init
.py", line 113, in get_cfg
check_dict_alignment(cfg, overrides)
File "C:\Users\MA\AppData\Local\Programs\Python\Python38\lib\site-packages\ultralytics\cfg_init_.py", line 184, in check_dict_alignment
raise SyntaxError(string + CLI_HELP_MSG) from e
SyntaxError: 'stream_buffer' is not a valid YOLO argument.

AttributeError: 'DXCamera' object has no attribute 'is_capturing'

E:\yolov8_aimbot-main>python run.py
PT Model loaded.
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT
Traceback (most recent call last):
File "E:\yolov8_aimbot-main\run.py", line 169, in
init()
File "C:\Python310\lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
File "E:\yolov8_aimbot-main\run.py", line 106, in init
frame = get_new_frame()
File "E:\yolov8_aimbot-main\logic\frame.py", line 24, in get_new_frame
dx = dxcam.create(device_idx=dxcam_monitor_id, output_idx=dxcam_gpu_id, output_color="BGR", max_buffer_len=dxcam_max_buffer_len)
File "C:\Python310\lib\site-packages\dxcam_init_.py", line 112, in create
return _factory.create(
File "C:\Python310\lib\site-packages\dxcam_init
.py", line 71, in create
camera = DXCamera(
File "C:\Python310\lib\site-packages\dxcam\dxcam.py", line 33, in init
self._duplicator: Duplicator = Duplicator(
File "", line 6, in init
File "C:\Python310\lib\site-packages\dxcam\core\duplicator.py", line 19, in post_init
output.output.DuplicateOutput(device.device, ctypes.byref(self.duplicator))
_ctypes.COMError: (-2005270524, '指定的设备接口或功能级别在此系统上不受支持。', (None, None, None, 0, None))
Exception ignored in: <function DXCamera.del at 0x00000206DB5B7EB0>
Traceback (most recent call last):
File "C:\Python310\lib\site-packages\dxcam\dxcam.py", line 243, in del
self.release()
File "C:\Python310\lib\site-packages\dxcam\dxcam.py", line 238, in release
self.stop()
File "C:\Python310\lib\site-packages\dxcam\dxcam.py", line 138, in stop
if self.is_capturing:
AttributeError: 'DXCamera' object has no attribute 'is_capturing'

I updated the graphics card driver but it still doesn't work. I don't know why

why pls help me!!!!!!!!!!!

Models Load "none"

Ultralytics YOLOv8.0.222 🚀 Python-3.11.6 torch-2.1.1+cu121 CUDA:0 (NVIDIA GeForce RTX 4090, 24564MiB)
Model summary (fused): 168 layers, 3007598 parameters, 0 gradients, 8.1 GFLOPs

PyTorch: starting from 'models\sunxds_0.2.1_nano_320.pt' with input shape (1, 3, 320, 320) BCHW and output shape(s) (1, 14, 2100) (5.9 MB)

ONNX: starting export with onnx 1.15.0 opset 17...

Is there steps to load models ?

Tha aimbot is open but no aimbot is active

I have followed the installation instructions and done every thing correctly, and when opening main.py file
it says
Model loaded. None
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT

in-game, I don't see any aimbot activity, playing on Destiny 2.

при запуски main не чего не происходит

Привет ещё раз я всё с темже обновлёная вами версия не запускается ошибок не каких не выдаёт она как бы запустилась но окно dxcam не появляется Старая версия работает исправно запускаются все три модели может что то не так с файлом main? Фаил checks запускал вроде не каких ошибок не выдаёт.
image
image
image

OSError: [WinError 126] Не найден указанный модуль. Error loading

OSError: [WinError 126] Не найден указанный модуль. Error loading "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\lib\cudnn_cnn_infer64_8.dll" or one of its dependencies.

У друга проблема, как ее решить? вроде был там тикет такой же, но не помогло когда др куду скачал, карта 1050 Ti

python main.py,Error

PS D:\yolo\yolov8-xhx> python main.py
Traceback (most recent call last):
File "D:\yolo\yolov8-xhx\main.py", line 2, in
from ultralytics import YOLO
File "G:\Python\Lib\site-packages\ultralytics_init_.py", line 5, in
from ultralytics.models import RTDETR, SAM, YOLO
File "G:\Python\Lib\site-packages\ultralytics\models_init_.py", line 3, in
from .rtdetr import RTDETR
File "G:\Python\Lib\site-packages\ultralytics\models\rtdetr_init_.py", line 3, in
from .model import RTDETR
File "G:\Python\Lib\site-packages\ultralytics\models\rtdetr\model.py", line 10, in
from ultralytics.engine.model import Model
File "G:\Python\Lib\site-packages\ultralytics\engine\model.py", line 8, in
from ultralytics.cfg import TASK2DATA, get_cfg, get_save_dir
File "G:\Python\Lib\site-packages\ultralytics\cfg_init_.py", line 10, in
from ultralytics.utils import (ASSETS, DEFAULT_CFG, DEFAULT_CFG_DICT, DEFAULT_CFG_PATH, LOGGER, RANK, ROOT, RUNS_DIR,
File "G:\Python\Lib\site-packages\ultralytics\utils_init_.py", line 19, in
import matplotlib.pyplot as plt
File "G:\Python\Lib\site-packages\matplotlib_init_.py", line 188, in
from . import _api, _version, cbook, _docstring, rcsetup
File "G:\Python\Lib\site-packages\matplotlib\cbook.py", line 32, in
from matplotlib import _api, _c_internal_utils
ImportError: cannot import name 'c_internal_utils' from partially initialized module 'matplotlib' (most likely due to a circular import) (G:\Python\Lib\site-packages\matplotlib_init.py)

Можно ли создать модель для PUBG?

Нашел модели на пабг, но они идут в фотографиях, ну в принципе да, я посмотрел что по ним и делается модель, но как саму модель по этим фото сделать? Больше всего это англоязычный обзоры, которые мне не помогают

My mouse will keep moving and won't be able to locate the character

My Python is version 3.9, and I have changed the "match cls" in the frame. py file to "if cls==0". Is this related?
My mouse will keep moving in one direction while holding down the right mouse button after starting "main. py", but it cannot accurately move to the character. How can I solve this problem

Вопрос по поводу карт Amd

Хотелось бы уточнить для незрячего,который только начинает знакомится с сие чудом,работает ли аим асист для видео карт амд,и если нет,есть ли какие то аналоги ,в частности интересует для проекта под названием warface,готов приобрести за нн-ую сумму,заранее благодарю за ответ.

NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend.

Model loaded. None
Traceback (most recent call last):
File "D:\23321\yolov8_aimbot-main\yolov8_aimbot-main\main.py", line 145, in
init()
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "D:\23321\yolov8_aimbot-main\yolov8_aimbot-main\main.py", line 113, in init
for frame in result:
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 35, in generator_context
response = gen.send(None)
^^^^^^^^^^^^^^
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 269, in stream_inference
self.results = self.postprocess(preds, im, im0s)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\models\yolo\detect\predict.py", line 25, in postprocess
preds = ops.non_max_suppression(preds,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\ops.py", line 238, in non_max_suppression
i = torchvision.ops.nms(boxes, scores, iou_thres) # NMS
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\torchvision\ops\boxes.py", line 41, in nms
return torch.ops.torchvision.nms(boxes, scores, iou_threshold)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\valor\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch_ops.py", line 692, in call
return self._op(*args, **kwargs or {})
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). If you are a Facebook employee using PyTorch on mobile, please visit https://fburl.com/ptmfixes for possible resolutions. 'torchvision::nms' is only available for these backends: [CPU, QuantizedCPU, BackendSelect, Python, FuncTorchDynamicLayerBackMode, Functionalize, Named, Conjugate, Negative, ZeroTensor, ADInplaceOrView, AutogradOther, AutogradCPU, AutogradCUDA, AutogradXLA, AutogradMPS, AutogradXPU, AutogradHPU, AutogradLazy, AutogradMeta, Tracer, AutocastCPU, AutocastCUDA, FuncTorchBatched, FuncTorchVmapMode, Batched, VmapMode, FuncTorchGradWrapper, PythonTLSSnapshot, FuncTorchDynamicLayerFrontMode, PreDispatch, PythonDispatcher].

CPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\cpu\nms_kernel.cpp:112 [kernel]
QuantizedCPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\quantized\cpu\qnms_kernel.cpp:124 [kernel]
BackendSelect: fallthrough registered at ..\aten\src\ATen\core\BackendSelectFallbackKernel.cpp:3 [backend fallback]
Python: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:153 [backend fallback]
FuncTorchDynamicLayerBackMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:498 [backend fallback]
Functionalize: registered at ..\aten\src\ATen\FunctionalizeFallbackKernel.cpp:290 [backend fallback]
Named: registered at ..\aten\src\ATen\core\NamedRegistrations.cpp:7 [backend fallback]
Conjugate: registered at ..\aten\src\ATen\ConjugateFallback.cpp:17 [backend fallback]
Negative: registered at ..\aten\src\ATen\native\NegateFallback.cpp:19 [backend fallback]
ZeroTensor: registered at ..\aten\src\ATen\ZeroTensorFallback.cpp:86 [backend fallback]
ADInplaceOrView: fallthrough registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:86 [backend fallback]
AutogradOther: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:53 [backend fallback]
AutogradCPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:57 [backend fallback]
AutogradCUDA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:65 [backend fallback]
AutogradXLA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:69 [backend fallback]
AutogradMPS: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:77 [backend fallback]
AutogradXPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:61 [backend fallback]
AutogradHPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:90 [backend fallback]
AutogradLazy: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:73 [backend fallback]
AutogradMeta: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:81 [backend fallback]
Tracer: registered at ..\torch\csrc\autograd\TraceTypeManual.cpp:296 [backend fallback]
AutocastCPU: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:382 [backend fallback]
AutocastCUDA: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:249 [backend fallback]
FuncTorchBatched: registered at ..\aten\src\ATen\functorch\LegacyBatchingRegistrations.cpp:710 [backend fallback]
FuncTorchVmapMode: fallthrough registered at ..\aten\src\ATen\functorch\VmapModeRegistrations.cpp:28 [backend fallback]
Batched: registered at ..\aten\src\ATen\LegacyBatchingRegistrations.cpp:1075 [backend fallback]
VmapMode: fallthrough registered at ..\aten\src\ATen\VmapModeRegistrations.cpp:33 [backend fallback]
FuncTorchGradWrapper: registered at ..\aten\src\ATen\functorch\TensorWrapper.cpp:203 [backend fallback]
PythonTLSSnapshot: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:161 [backend fallback]
FuncTorchDynamicLayerFrontMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:494 [backend fallback]
PreDispatch: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:165 [backend fallback]
PythonDispatcher: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:157 [backend fallback]

вв

Вопрос, как запустить Onnx модель на амд ? Чтоб проканало

How to continue training the model?

Please tell me how to continue training the model. I found that the recognition rate in Battlefield 5 is only 30%. Do I need to take 1,000 screenshots first? Then how to label it and how to continue training?

error in main.py line 102

after getting everything setup and all verified ok with the checks.py when trying to run the main.py users will be met with the following error "SyntaxError: 'show_boxes' is not a valid YOLO argument. Similar arguments are i.e. ['show_labels=True', 'boxes=True', 'show_conf=True']." to fix this i went to line 102* "show_boxes=False," and simply erased the "show_" and it works fine sorry if this wasn't worth opening an issue just thought you should be made aware.

NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend.

C:\Users\USER\Desktop\yolov8_aimbot-main>python main.py.
Model loaded. None
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT
Traceback (most recent call last):
File "C:\Users\USER\Desktop\yolov8_aimbot-main\main.py", line 161, in
init()
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\Desktop\yolov8_aimbot-main\main.py", line 127, in init
for frame in result:
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 35, in generator_context
response = gen.send(None)
^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 269, in stream_inference
self.results = self.postprocess(preds, im, im0s)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\models\yolo\detect\predict.py", line 25, in postprocess
preds = ops.non_max_suppression(preds,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\ops.py", line 238, in non_max_suppression
i = torchvision.ops.nms(boxes, scores, iou_thres) # NMS
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\torchvision\ops\boxes.py", line 41, in nms
return torch.ops.torchvision.nms(boxes, scores, iou_threshold)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\USER\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch_ops.py", line 692, in call
return self._op(*args, **kwargs or {})
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). If you are a Facebook employee using PyTorch on mobile, please visit https://fburl.com/ptmfixes for possible resolutions. 'torchvision::nms' is only available for these backends: [CPU, QuantizedCPU, BackendSelect, Python, FuncTorchDynamicLayerBackMode, Functionalize, Named, Conjugate, Negative, ZeroTensor, ADInplaceOrView, AutogradOther, AutogradCPU, AutogradCUDA, AutogradXLA, AutogradMPS, AutogradXPU, AutogradHPU, AutogradLazy, AutogradMeta, Tracer, AutocastCPU, AutocastCUDA, FuncTorchBatched, FuncTorchVmapMode, Batched, VmapMode, FuncTorchGradWrapper, PythonTLSSnapshot, FuncTorchDynamicLayerFrontMode, PreDispatch, PythonDispatcher].

CPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\cpu\nms_kernel.cpp:112 [kernel]
QuantizedCPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\quantized\cpu\qnms_kernel.cpp:124 [kernel]
BackendSelect: fallthrough registered at ..\aten\src\ATen\core\BackendSelectFallbackKernel.cpp:3 [backend fallback]
Python: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:153 [backend fallback]
FuncTorchDynamicLayerBackMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:498 [backend fallback]
Functionalize: registered at ..\aten\src\ATen\FunctionalizeFallbackKernel.cpp:290 [backend fallback]
Named: registered at ..\aten\src\ATen\core\NamedRegistrations.cpp:7 [backend fallback]
Conjugate: registered at ..\aten\src\ATen\ConjugateFallback.cpp:17 [backend fallback]
Negative: registered at ..\aten\src\ATen\native\NegateFallback.cpp:19 [backend fallback]
ZeroTensor: registered at ..\aten\src\ATen\ZeroTensorFallback.cpp:86 [backend fallback]
ADInplaceOrView: fallthrough registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:86 [backend fallback]
AutogradOther: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:53 [backend fallback]
AutogradCPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:57 [backend fallback]
AutogradCUDA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:65 [backend fallback]
AutogradXLA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:69 [backend fallback]
AutogradMPS: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:77 [backend fallback]
AutogradXPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:61 [backend fallback]
AutogradHPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:90 [backend fallback]
AutogradLazy: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:73 [backend fallback]
AutogradMeta: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:81 [backend fallback]
Tracer: registered at ..\torch\csrc\autograd\TraceTypeManual.cpp:296 [backend fallback]
AutocastCPU: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:382 [backend fallback]
AutocastCUDA: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:249 [backend fallback]
FuncTorchBatched: registered at ..\aten\src\ATen\functorch\LegacyBatchingRegistrations.cpp:710 [backend fallback]
FuncTorchVmapMode: fallthrough registered at ..\aten\src\ATen\functorch\VmapModeRegistrations.cpp:28 [backend fallback]
Batched: registered at ..\aten\src\ATen\LegacyBatchingRegistrations.cpp:1075 [backend fallback]
VmapMode: fallthrough registered at ..\aten\src\ATen\VmapModeRegistrations.cpp:33 [backend fallback]
FuncTorchGradWrapper: registered at ..\aten\src\ATen\functorch\TensorWrapper.cpp:203 [backend fallback]
PythonTLSSnapshot: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:161 [backend fallback]
FuncTorchDynamicLayerFrontMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:494 [backend fallback]
PreDispatch: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:165 [backend fallback]
PythonDispatcher: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:157 [backend fallback]

main.py torch.cuda.is_available(): False

Здравствуйте, делаю все как в гайде.ru указано, не работает. После ввода python main.py выдает следующее:C:\Program Files\yolov8_aimbot-main>python main.py
Model loaded. None
Ultralytics YOLOv8.0.221 🚀 Python-3.11.6 torch-2.1.1+cpu
Traceback (most recent call last):
File "C:\Program Files\yolov8_aimbot-main\main.py", line 145, in
init()
File "C:\Users\rx 580 hlam no drive\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Program Files\yolov8_aimbot-main\main.py", line 86, in init
result = model.predict(
^^^^^^^^^^^^^^
File "C:\Users\rx 580 hlam no drive\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\model.py", line 232, in predict
self.predictor.setup_model(model=self.model, verbose=is_cli)
File "C:\Users\rx 580 hlam no drive\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 318, in setup_model
device=select_device(self.args.device, verbose=verbose),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\rx 580 hlam no drive\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\torch_utils.py", line 119, in select_device
raise ValueError(f"Invalid CUDA 'device={device}' requested."
ValueError: Invalid CUDA 'device=0' requested. Use 'device=cpu' or pass valid CUDA device(s) if available, i.e. 'device=0' or 'device=0,1,2,3' for Multi-GPU.

torch.cuda.is_available(): False
torch.cuda.device_count(): 0
os.environ['CUDA_VISIBLE_DEVICES']: None
See https://pytorch.org/get-started/locally/ for up-to-date torch install instructions if no CUDA devices are seen by torch.

Видео карта 1050. Если что то еще нужно скинуть, я скину. Переустановка pytrch не помогает
помогите пажалуйста

RTX 4090 ? and issue TensorRT: export failure ❌ 0.6s: Could not find: cudnn64_8.dll.

PyTorch: starting from 'models\sunxds_0.2.1_nano_320.pt' with input shape (1, 3, 320, 320) BCHW and output shape(s) (1, 14, 2100) (5.9 MB) TensorRT: export failure ❌ 0.6s: Could not find: cudnn64_8.dll. Is it on your PATH? Note: Paths searched were: ['C:\\Users\\vakut\\AppData\\Local\\Programs\\Python\\Python311\\Lib\\site-packages\\cv2\\../../x64/vc14/bin', 'C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.1\\bin', 'C:\\Program Files\\NVIDIA GPU Computing Toolkit\\CUDA\\v12.1\\libnvvp', 'C:\\Windows\\system32', 'C:\\Windows', 'C:\\Windows\\System32\\Wbem', 'C:\\Windows\\System32\\WindowsPowerShell\\v1.0\\', 'C:\\Windows\\System32\\OpenSSH\\', 'C:\\Program Files\\NVIDIA Corporation\\NVIDIA NvDLISR', 'C:\\Program Files (x86)\\NVIDIA Corporation\\PhysX\\Common', 'C:\\Program Files\\NVIDIA Corporation\\Nsight Compute 2023.1.0\\', 'C:\\Users\\vakut\\OneDrive\\Desktop\\yolov8_aimbot-main\\TensorRT-8.6.1.6\\TensorRT-8.6.1.6\\lib', 'C:\\Users\\vakut\\AppData\\Local\\Programs\\Python\\Python311\\Scripts\\', 'C:\\Users\\vakut\\AppData\\Local\\Programs\\Python\\Python311\\', 'C:\\Users\\vakut\\AppData\\Local\\Microsoft\\WindowsApps', ''] Traceback (most recent call last): File "<frozen runpy>", line 198, in _run_module_as_main File "<frozen runpy>", line 88, in _run_code File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Scripts\yolo.exe\__main__.py", line 7, in <module> File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\cfg\__init__.py", line 448, in entrypoint getattr(model, mode)(**overrides) # default args from model ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\model.py", line 310, in export return Exporter(overrides=args, _callbacks=self.callbacks)(model=self.model) ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils\_contextlib.py", line 115, in decorate_context return func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 252, in __call__ f[1], _ = self.export_engine() ^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 122, in outer_func raise e File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 117, in outer_func f, model = inner_func(*args, **kwargs) ^^^^^^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 578, in export_engine import tensorrt as trt # noqa ^^^^^^^^^^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\tensorrt\__init__.py", line 127, in <module> ctypes.CDLL(find_lib(lib)) ^^^^^^^^^^^^^ File "C:\Users\vakut\AppData\Local\Programs\Python\Python311\Lib\site-packages\tensorrt\__init__.py", line 81, in find_lib raise FileNotFoundError( FileNotFoundError: Could not find: cudnn64_8.dll. Is it on your PATH?

Something wrong with a path could it be issue with the Graphics card ?

Ошибка экспорта

Когда установлен модуль «onnx» ничего не происходит:
Screenshot_9

Когда прописываю Pip uninstall onnx, как было в рекомендацих
Screenshot_10

Не могу выключить нейронку

Такая проблема, не запускаю cmd, захожу поиграть просто, но все равно нейронка работает,до наводит на соперника хотя я ее не включал, как ее можно выключить?
Помогите пожалуйста

настройка

Кто то может скинуть настройки чтобы аимило четко не трясся прицел в стороны нвидиа 1050 на 4 гб стоит можете скинуть сетап ?

Помогите))

C:\Users\Windows 10\Desktop\yolov8_aimbot-main>python main.py
Model loaded. None
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT
Ultralytics YOLOv8.0.222 🚀 Python-3.11.6 torch-2.1.1+cpu
Traceback (most recent call last):
File "C:\Users\Windows 10\Desktop\yolov8_aimbot-main\main.py", line 160, in
init()
File "C:\Users\Windows 10\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\Windows 10\Desktop\yolov8_aimbot-main\main.py", line 101, in init
result = model.predict(
^^^^^^^^^^^^^^
File "C:\Users\Windows 10\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\model.py", line 232, in predict
self.predictor.setup_model(model=self.model, verbose=is_cli)
File "C:\Users\Windows 10\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 318, in setup_model
device=select_device(self.args.device, verbose=verbose),
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\Windows 10\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\torch_utils.py", line 119, in select_device
raise ValueError(f"Invalid CUDA 'device={device}' requested."
ValueError: Invalid CUDA 'device=0' requested. Use 'device=cpu' or pass valid CUDA device(s) if available, i.e. 'device=0' or 'device=0,1,2,3' for Multi-GPU.

torch.cuda.is_available(): False
torch.cuda.device_count(): 0
os.environ['CUDA_VISIBLE_DEVICES']: None
See https://pytorch.org/get-started/locally/ for up-to-date torch install instructions if no CUDA devices are seen by torch.

разрешение

если у меня 4:3 играю в варфейс, в конфиге AI image size надо увеличивать или нет? модельки же у меня более растянутые

NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend.

в общем все сделал все работает но выдает такое и не наводится в игре что делать ?
Microsoft Windows [Version 10.0.19045.3693]
(c) Корпорация Майкрософт (Microsoft Corporation). Все права защищены.

C:\Users\igoty\Desktop\yolov8_aimbot-main>python main.py
Model loaded. None
Aimbot is started. Enjoy!
[Right mouse button] - Aiming at the target
[F2] - EXIT
Traceback (most recent call last):
File "C:\Users\igoty\Desktop\yolov8_aimbot-main\main.py", line 161, in
init()
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\igoty\Desktop\yolov8_aimbot-main\main.py", line 127, in init
for frame in result:
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 35, in generator_context
response = gen.send(None)
^^^^^^^^^^^^^^
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 269, in stream_inference
self.results = self.postprocess(preds, im, im0s)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\models\yolo\detect\predict.py", line 25, in postprocess
preds = ops.non_max_suppression(preds,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\ops.py", line 238, in non_max_suppression
i = torchvision.ops.nms(boxes, scores, iou_thres) # NMS
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\torchvision\ops\boxes.py", line 41, in nms
return torch.ops.torchvision.nms(boxes, scores, iou_threshold)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\igoty\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch_ops.py", line 692, in call
return self._op(*args, **kwargs or {})
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). If you are a Facebook employee using PyTorch on mobile, please visit https://fburl.com/ptmfixes for possible resolutions. 'torchvision::nms' is only available for these backends: [CPU, QuantizedCPU, BackendSelect, Python, FuncTorchDynamicLayerBackMode, Functionalize, Named, Conjugate, Negative, ZeroTensor, ADInplaceOrView, AutogradOther, AutogradCPU, AutogradCUDA, AutogradXLA, AutogradMPS, AutogradXPU, AutogradHPU, AutogradLazy, AutogradMeta, Tracer, AutocastCPU, AutocastCUDA, FuncTorchBatched, FuncTorchVmapMode, Batched, VmapMode, FuncTorchGradWrapper, PythonTLSSnapshot, FuncTorchDynamicLayerFrontMode, PreDispatch, PythonDispatcher].

CPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\cpu\nms_kernel.cpp:112 [kernel]
QuantizedCPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\quantized\cpu\qnms_kernel.cpp:124 [kernel]
BackendSelect: fallthrough registered at ..\aten\src\ATen\core\BackendSelectFallbackKernel.cpp:3 [backend fallback]
Python: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:153 [backend fallback]
FuncTorchDynamicLayerBackMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:498 [backend fallback]
Functionalize: registered at ..\aten\src\ATen\FunctionalizeFallbackKernel.cpp:290 [backend fallback]
Named: registered at ..\aten\src\ATen\core\NamedRegistrations.cpp:7 [backend fallback]
Conjugate: registered at ..\aten\src\ATen\ConjugateFallback.cpp:17 [backend fallback]
Negative: registered at ..\aten\src\ATen\native\NegateFallback.cpp:19 [backend fallback]
ZeroTensor: registered at ..\aten\src\ATen\ZeroTensorFallback.cpp:86 [backend fallback]
ADInplaceOrView: fallthrough registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:86 [backend fallback]
AutogradOther: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:53 [backend fallback]
AutogradCPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:57 [backend fallback]
AutogradCUDA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:65 [backend fallback]
AutogradXLA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:69 [backend fallback]
AutogradMPS: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:77 [backend fallback]
AutogradXPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:61 [backend fallback]
AutogradHPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:90 [backend fallback]
AutogradLazy: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:73 [backend fallback]
AutogradMeta: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:81 [backend fallback]
Tracer: registered at ..\torch\csrc\autograd\TraceTypeManual.cpp:296 [backend fallback]
AutocastCPU: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:382 [backend fallback]
AutocastCUDA: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:249 [backend fallback]
FuncTorchBatched: registered at ..\aten\src\ATen\functorch\LegacyBatchingRegistrations.cpp:710 [backend fallback]
FuncTorchVmapMode: fallthrough registered at ..\aten\src\ATen\functorch\VmapModeRegistrations.cpp:28 [backend fallback]
Batched: registered at ..\aten\src\ATen\LegacyBatchingRegistrations.cpp:1075 [backend fallback]
VmapMode: fallthrough registered at ..\aten\src\ATen\VmapModeRegistrations.cpp:33 [backend fallback]
FuncTorchGradWrapper: registered at ..\aten\src\ATen\functorch\TensorWrapper.cpp:203 [backend fallback]
PythonTLSSnapshot: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:161 [backend fallback]
FuncTorchDynamicLayerFrontMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:494 [backend fallback]
PreDispatch: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:165 [backend fallback]
PythonDispatcher: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:157 [backend fallback]

NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend.

C:\yolov8_aimbot-main>python main.py
Model loaded. None
Traceback (most recent call last):
File "C:\yolov8_aimbot-main\main.py", line 145, in
init()
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\yolov8_aimbot-main\main.py", line 113, in init
for frame in result:
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 35, in generator_context
response = gen.send(None)
^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\predictor.py", line 269, in stream_inference
self.results = self.postprocess(preds, im, im0s)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\models\yolo\detect\predict.py", line 25, in postprocess
preds = ops.non_max_suppression(preds,
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\ops.py", line 238, in non_max_suppression
i = torchvision.ops.nms(boxes, scores, iou_thres) # NMS
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\torchvision\ops\boxes.py", line 41, in nms
return torch.ops.torchvision.nms(boxes, scores, iou_threshold)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch_ops.py", line 692, in call
return self._op(*args, **kwargs or {})
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
NotImplementedError: Could not run 'torchvision::nms' with arguments from the 'CUDA' backend. This could be because the operator doesn't exist for this backend, or was omitted during the selective/custom build process (if using custom build). If you are a Facebook employee using PyTorch on mobile, please visit https://fburl.com/ptmfixes for possible resolutions. 'torchvision::nms' is only available for these backends: [CPU, QuantizedCPU, BackendSelect, Python, FuncTorchDynamicLayerBackMode, Functionalize, Named, Conjugate, Negative, ZeroTensor, ADInplaceOrView, AutogradOther, AutogradCPU, AutogradCUDA, AutogradXLA, AutogradMPS, AutogradXPU, AutogradHPU, AutogradLazy, AutogradMeta, Tracer, AutocastCPU, AutocastCUDA, FuncTorchBatched, FuncTorchVmapMode, Batched, VmapMode, FuncTorchGradWrapper, PythonTLSSnapshot, FuncTorchDynamicLayerFrontMode, PreDispatch, PythonDispatcher].

CPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\cpu\nms_kernel.cpp:112 [kernel]
QuantizedCPU: registered at C:\actions-runner_work\vision\vision\pytorch\vision\torchvision\csrc\ops\quantized\cpu\qnms_kernel.cpp:124 [kernel]
BackendSelect: fallthrough registered at ..\aten\src\ATen\core\BackendSelectFallbackKernel.cpp:3 [backend fallback]
Python: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:153 [backend fallback]
FuncTorchDynamicLayerBackMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:498 [backend fallback]
Functionalize: registered at ..\aten\src\ATen\FunctionalizeFallbackKernel.cpp:290 [backend fallback]
Named: registered at ..\aten\src\ATen\core\NamedRegistrations.cpp:7 [backend fallback]
Conjugate: registered at ..\aten\src\ATen\ConjugateFallback.cpp:17 [backend fallback]
Negative: registered at ..\aten\src\ATen\native\NegateFallback.cpp:19 [backend fallback]
ZeroTensor: registered at ..\aten\src\ATen\ZeroTensorFallback.cpp:86 [backend fallback]
ADInplaceOrView: fallthrough registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:86 [backend fallback]
AutogradOther: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:53 [backend fallback]
AutogradCPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:57 [backend fallback]
AutogradCUDA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:65 [backend fallback]
AutogradXLA: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:69 [backend fallback]
AutogradMPS: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:77 [backend fallback]
AutogradXPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:61 [backend fallback]
AutogradHPU: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:90 [backend fallback]
AutogradLazy: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:73 [backend fallback]
AutogradMeta: registered at ..\aten\src\ATen\core\VariableFallbackKernel.cpp:81 [backend fallback]
Tracer: registered at ..\torch\csrc\autograd\TraceTypeManual.cpp:296 [backend fallback]
AutocastCPU: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:382 [backend fallback]
AutocastCUDA: fallthrough registered at ..\aten\src\ATen\autocast_mode.cpp:249 [backend fallback]
FuncTorchBatched: registered at ..\aten\src\ATen\functorch\LegacyBatchingRegistrations.cpp:710 [backend fallback]
FuncTorchVmapMode: fallthrough registered at ..\aten\src\ATen\functorch\VmapModeRegistrations.cpp:28 [backend fallback]
Batched: registered at ..\aten\src\ATen\LegacyBatchingRegistrations.cpp:1075 [backend fallback]
VmapMode: fallthrough registered at ..\aten\src\ATen\VmapModeRegistrations.cpp:33 [backend fallback]
FuncTorchGradWrapper: registered at ..\aten\src\ATen\functorch\TensorWrapper.cpp:203 [backend fallback]
PythonTLSSnapshot: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:161 [backend fallback]
FuncTorchDynamicLayerFrontMode: registered at ..\aten\src\ATen\functorch\DynamicLayer.cpp:494 [backend fallback]
PreDispatch: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:165 [backend fallback]
PythonDispatcher: registered at ..\aten\src\ATen\core\PythonFallbackKernel.cpp:157 [backend fallback]

C:\yolov8_aimbot-main>yolo export model="models/sunxds_0.2.1_nano_320.pt" format=engine half=true device=0 workspace=8 imgsz=320c
Ultralytics YOLOv8.0.221 🚀 Python-3.11.6 torch-2.1.1+cu121 CUDA:0 (NVIDIA GeForce GTX 1650 SUPER, 4096MiB)
Traceback (most recent call last):
File "", line 198, in run_module_as_main
File "", line 88, in run_code
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Scripts\yolo.exe_main
.py", line 7, in
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\cfg_init
.py", line 448, in entrypoint
getattr(model, mode)(**overrides) # default args from model
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\model.py", line 310, in export
return Exporter(overrides=args, _callbacks=self.callbacks)(model=self.model)
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\torch\utils_contextlib.py", line 115, in decorate_context
return func(*args, **kwargs)
^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\engine\exporter.py", line 180, in call
self.imgsz = check_imgsz(self.args.imgsz, stride=model.stride, min_dim=2) # check image size
^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
File "C:\Users\dsd-2\AppData\Local\Programs\Python\Python311\Lib\site-packages\ultralytics\utils\checks.py", line 124, in check_imgsz
raise TypeError(f"'imgsz={imgsz}' is of invalid type {type(imgsz).name}. "
TypeError: 'imgsz=320c' is of invalid type str. Valid imgsz types are int i.e. 'imgsz=640' or list i.e. 'imgsz=[640,640]'

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.