Linux ubuntu 6.8.0-90-generic #91-Ubuntu SMP PREEMPT_DYNAMIC Tue Nov 18 14:14:30 UTC 2025 x86_64
nginx/1.24.0
: 67.217.245.49 | : 216.73.216.153
Cant Read [ /etc/named.conf ]
8.3.6
www-data
Bypass.pw
Terminal
AUTO ROOT
Adminer
Backdoor Destroyer
Linux Exploit
Lock Shell
Lock File
Create User
CREATE RDP
PHP Mailer
BACKCONNECT
UNLOCK SHELL
HASH IDENTIFIER
Backdoor Scanner
Backdoor Create
Alfa Webshell
CPANEL RESET
CREATE WP USER
README
+ Create Folder
+ Create File
/
usr /
lib /
pcp /
pmdas /
nvidia /
[ HOME SHELL ]
Name
Size
Permission
Action
Install
728
B
-rwxr-xr-x
README
215
B
-rw-r--r--
Remove
1.03
KB
-rwxr-xr-x
domain.h
57
B
-rw-r--r--
help
6.27
KB
-rw-r--r--
pmda_nvidia.so
33.04
KB
-rw-r--r--
pmdanvidia
32.67
KB
-rwxr-xr-x
pmns
1.96
KB
-rw-r--r--
root
115
B
-rw-r--r--
Delete
Unzip
Zip
${this.title}
Close
Code Editor : help
# # Copyright (c) 2014,2019,2021 Red Hat. # Copyright (c) 2000-2004 Silicon Graphics, Inc. All Rights Reserved. # # This program is free software; you can redistribute it and/or modify it # under the terms of the GNU General Public License as published by the # Free Software Foundation; either version 2 of the License, or (at your # option) any later version. # # This program is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY # or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License # for more details. # # NVIDIA PMDA help file in the ASCII format # # lines beginning with a # are ignored # lines beginning @ introduce a new entry of the form # @ metric_name oneline-text # help text goes # here over multiple lines # ... # # the metric_name is decoded against the default PMNS -- as a special case, # a name of the form NNN.MM (for numeric NNN and MM) is interpreted as an # instance domain identification, and the text describes the instance domain # # blank lines before the @ line are ignored # @ 120.0 NVIDIA graphics cards installed in this system @ 120.1 processes mapped to NVIDIA graphics cards in use @ 120.2 set of processes using NVIDIA graphics card resources @ nvidia.numcards Number of Graphics Cards The number of NVIDIA Graphics cards installed in this system @ nvidia.gpuid GPU ID Zero indexed id of this NVIDIA card @ nvidia.cardname GPU Name The name of the graphics card @ nvidia.carduuid GPU Universally Unique ID The Universally Unique ID of the graphics card @ nvidia.busid Card Bus ID The Bus ID as reported by the NVIDIA tools, not lspci @ nvidia.temperature The temperature of the card The temperature of the GPU on the NVIDIA card in degrees celsius. @ nvidia.fanspeed Fanspeed Speed of the GPU fan as a percentage of the maximum @ nvidia.perfstate NVIDIA performance state The PX performance state as reported from NVML. Value is an integer which should range from 0 (maximum performance) to 15 (minimum). If the state is unknown the reported value will be 32, however. @ nvidia.gpuactive Percentage of GPU utilization Percentage of time over the past sample period during which one or more kernels was executing on the GPU. @ nvidia.memactive Percentage of time spent accessing memory Percent of time over the past sample period during which global (device) memory was being read or written. This metric shows if the memory is actively being accessed, and is not correlated to storage amount used. @ nvidia.memused Allocated frame buffer memory Amount of GPU FB memory that has currently been allocated, in bytes. Note that the driver/GPU always sets aside a small amount of memory for bookkeeping. @ nvidia.memtotal Total frame buffer memory available The total amount of GPU FB memory available on the card, in bytes. @ nvidia.memfree Unallocated frame buffer memory Amount of GPU FB memory that is not currently allocated, in bytes. @ nvidia.energy Total GPU energy consumption Total energy consumption for each GPU, in millijoules (mJ). @ nvidia.power Total GPU power consumption Current power usage for this GPU and its associated circuitry, in milliwatts. @ nvidia.nprocs Count of processes using each GPU @ nvidia.samples Count of value refreshes for each GPU @ nvidia.gpuutilaccum Cumulative counter of GPU utilization for each GPU @ nvidia.memutilaccum Cumulative counter of memory utilization for each GPU @ nvidia.memusedaccum Cumulative counter of used memory for each GPU @ nvidia.proc.samples Number of times process stats have been sampled @ nvidia.proc.memused Memory in use by each process accessing each card @ nvidia.proc.memaccum Accumulated memory used by processes accessing each card @ nvidia.proc.gpuactive Graphics card utilization by processes accessing each card @ nvidia.proc.memactive Memory access utilization by processes accessing each card @ nvidia.proc.time Milliseconds spent by processes accessing each graphics card @ nvidia.proc.running Indicates whether each process accessing each card is still running @ nvidia.proc.all.memused Memory in use by each process across all cards @ nvidia.proc.all.gpuactive Graphics card utilization by processes accessing all cards @ nvidia.proc.all.memactive Memory utilization by processes accessing all cards @ nvidia.proc.all.time Milliseconds spent by processes accessing all graphics cards @ nvidia.proc.all.running Indicates whether each process is still running @ nvidia.proc.all.gpulist Bitmask of GPUs used by each process @ nvidia.proc.all.ngpus Count of GPUs used by each process @ nvidia.proc.all.samples Number of times process stats across all cards have been sampled @ nvidia.proc.all.memaccum Cumulative count of memory used by each process across all GPUs @ nvidia.proc.compute.memused Compute memory in use by processes across all cards @ nvidia.proc.compute.gpuactive Compute GPU utilization by processes accessing all cards @ nvidia.proc.compute.memactive Compute memory utilization by processes accessing all cards @ nvidia.proc.compute.time Compute time spent by processes accessing all cards @ nvidia.proc.compute.running Indicates whether each process using compute is still running @ nvidia.proc.compute.gpulist Bitmask of GPUs used for compute by each process @ nvidia.proc.compute.ngpus Count of GPUs used for compute by each process @ nvidia.proc.compute.samples Number of metric refreshes for processes using compute across all cards @ nvidia.proc.compute.memaccum Cumulative count of compute memory used by each process across all GPUs @ nvidia.proc.graphics.memused Graphics memory in use by processes across all cards @ nvidia.proc.graphics.gpuactive Graphics GPU utilization by processes accessing all cards @ nvidia.proc.graphics.memactive Graphics memory utilization by processes accessing all cards @ nvidia.proc.graphics.time Graphics time spent by processes accessing all cards @ nvidia.proc.graphics.running Indicates whether each process using graphics is still running @ nvidia.proc.graphics.gpulist Bitmask of GPUs used for graphics by each process @ nvidia.proc.graphics.ngpus Count of GPUs used for graphics by each process @ nvidia.proc.graphics.samples Number of metric refreshes for processes using graphics across all cards @ nvidia.proc.graphics.memaccum Cumulative count of graphics memory used by each process across all GPUs
Close