diff --git a/doc/variables.xml b/doc/variables.xml
index c69f13f8..1615ec00 100644
--- a/doc/variables.xml
+++ b/doc/variables.xml
@@ -2781,10 +2781,14 @@
+
Nvidia graphics card information via the XNVCtrl
library.
-
+
+ GPU_ID: Optional parameter to choose the GPU to be used as 0,1,2,3,.. Default parameter is 0
+
+
Possible arguments: (Temperatures are
printed as float, all other values as integer. Bracketed
arguments are aliases)
@@ -2925,8 +2929,12 @@
imagequality
+
+ modelname
+
+
-
+
@@ -2935,12 +2943,16 @@
+
Same as nvidia, except it draws its output in a
horizontal bar. The height and width parameters are optional,
and default to the default_bar_height and default_bar_width
config settings, respectively.
-
+
+ GPU_ID: Optional parameter to choose the GPU to be used as 0,1,2,3,.. Default parameter is 0
+
+
Note the following arguments are incompatible:
@@ -2982,7 +2994,7 @@
fanspeed
-
+
@@ -2991,15 +3003,19 @@
+
Same as nvidiabar, except a round gauge
(much like a vehicle speedometer). The height
and width parameters are optional, and default to the
default_gauge_height and default_gauge_width config
settings, respectively.
-
+
+ GPU_ID: Optional parameter to choose the GPU to be used as 0,1,2,3,.. Default parameter is 0
+
+
For possible arguments see nvidia and nvidiabar.
-
+
@@ -3013,17 +3029,21 @@
+
Same as nvidiabar, except a horizontally
scrolling graph with values from 0-100 plotted on the
vertical axis. The height and width parameters are
optional, and default to the default_graph_height and
default_graph_width config settings, respectively.
-
+
+ GPU_ID: NOT optional. This parameter allows to choose the GPU to be used as 0,1,2,3,..
+
+
For possible arguments see nvidia and nvidiabar. To learn more
about the -t -l and gradient color options,
see execgraph.
-
+
diff --git a/src/nvidia.cc b/src/nvidia.cc
index d5871356..6c705f0c 100644
--- a/src/nvidia.cc
+++ b/src/nvidia.cc
@@ -35,8 +35,6 @@
* Fonic
*
* TODO:
- * - Add third argument to module to allow querying multiple GPUs/fans etc.,
- * e.g. ${nvidia gputemp 2}, ${nvidia fanlevel 1}
* - Move decoding of GPU/MEM freqs to print_nvidia_value() using QUERY_SPECIAL
* so that all quirks are located there
* - Implement nvs->print_type to allow control over how the value is printed
@@ -44,24 +42,45 @@
*
* Showcase (conky.conf):
* --==| NVIDIA | ==--
- * GPU ${nvidia gpufreq}MHz (${nvidia gpufreqmin}-${nvidia gpufreqmax}MHz)
- * MEM ${nvidia memfreq}MHz (${nvidia memfreqmin}-${nvidia memfreqmax}MHz)
- * MTR ${nvidia mtrfreq}MHz (${nvidia mtrfreqmin}-${nvidia mtrfreqmax}MHz)
- * PERF Level ${nvidia perflevel} (${nvidia perflevelmin}-${nvidia perflevelmax}), Mode: ${nvidia perfmode}
- * VRAM ${nvidia memutil}% (${nvidia memused}MB/${nvidia memtotal}MB)
- * LOAD GPU ${nvidia gpuutil}%, RAM ${nvidia membwutil}%, VIDEO ${nvidia videoutil}%, PCIe ${nvidia pcieutil}%
- * TEMP GPU ${nvidia gputemp}°C (${nvidia gputempthreshold}°C max.), SYS ${nvidia ambienttemp}°C
- * FAN ${nvidia fanspeed} RPM (${nvidia fanlevel}%)
+ * GPU ${nvidia gpufreq [gpu_id]}MHz (${nvidia gpufreqmin [gpu_id]}-${nvidia gpufreqmax [gpu_id]}MHz)
+ * MEM ${nvidia memfreq [gpu_id]}MHz (${nvidia memfreqmin [gpu_id]}-${nvidia memfreqmax [gpu_id]}MHz)
+ * MTR ${nvidia mtrfreq [gpu_id]}MHz (${nvidia mtrfreqmin [gpu_id]}-${nvidia mtrfreqmax [gpu_id]}MHz)
+ * PERF Level ${nvidia perflevel [gpu_id]} (${nvidia perflevelmin [gpu_id]}-${nvidia perflevelmax [gpu_id]}), Mode: ${nvidia perfmode [gpu_id]}
+ * VRAM ${nvidia memutil [gpu_id]}% (${nvidia memused [gpu_id]}MB/${nvidia memtotal [gpu_id]}MB)
+ * LOAD GPU ${nvidia gpuutil [gpu_id]}%, RAM ${nvidia membwutil [gpu_id]}%, VIDEO ${nvidia videoutil [gpu_id]}%, PCIe ${nvidia pcieutil [gpu_id]}%
+ * TEMP GPU ${nvidia gputemp [gpu_id]}°C (${nvidia gputempthreshold [gpu_id]}°C max.), SYS ${nvidia ambienttemp [gpu_id]}°C
+ * FAN ${nvidia fanspeed [gpu_id]} RPM (${nvidia fanlevel [gpu_id]}%)
+ *
+ * miscelaneas:
+ * OPENGL ${nvidia imagequality [gpu_id]}
+ * GPU ${nvidia modelname [gpu_id]}
*
* --==| NVIDIA Bars |==--
- * LOAD ${nvidiabar gpuutil}
- * VRAM ${nvidiabar memutil}
- * RAM ${nvidiabar membwutil}
- * VIDEO ${nvidiabar videoutil}
- * PCIe ${nvidiabar pcieutil}
- * Fan ${nvidiabar fanlevel}
- * TEMP ${nvidiabar gputemp}
+ * LOAD ${nvidiabar [height][,width] gpuutil [gpu_id]}
+ * VRAM ${nvidiabar [height][,width] memutil [gpu_id]}
+ * RAM ${nvidiabar [height][,width] membwutil [gpu_id]}
+ * VIDEO ${nvidiabar [height][,width] videoutil [gpu_id]}
+ * PCIe ${nvidiabar [height][,width] pcieutil [gpu_id]}
+ * Fan ${nvidiabar [height][,width] fanlevel [gpu_id]}
+ * TEMP ${nvidiabar [height][,width] gputemp [gpu_id]}
*
+ * --==| NVIDIA Gauge |==--
+ * LOAD ${nvidiagauge [height][,width] gpuutil [gpu_id]}
+ * VRAM ${nvidiagauge [height][,width] memutil [gpu_id]}
+ * RAM ${nvidiagauge [height][,width] membwutil [gpu_id]}
+ * VIDEO ${nvidiagauge [height][,width] videoutil [gpu_id]}
+ * PCIe ${nvidiagauge [height][,width] pcieutil [gpu_id]}
+ * Fan ${nvidiagauge [height][,width] fanlevel [gpu_id]}
+ * TEMP ${nvidiagauge [height][,width] gputemp [gpu_id]}
+ *
+ * --==| NVIDIA Graph |==-- (gpu_id is not optional in this case)
+ * LOAD ${nvidiagraph gpuutil [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * VRAM ${nvidiagraph memutil [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * RAM ${nvidiagraph membwutil [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * VIDEO ${nvidiagraph videoutil [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * PCIe ${nvidiagraph pcieutil [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * Fan ${nvidiagraph fanlevel [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
+ * TEMP ${nvidiagraph gputemp [height][,width] [gradient color 1] [gradient color 2] [scale] [-t] [-l] gpu_id}
*/
@@ -126,7 +145,8 @@ const char* translate_module_argument[] = {
"fanspeed", // Fan/cooler
"fanlevel",
- "imagequality" // Miscellaneous
+ "imagequality", // Miscellaneous
+ "modelname"
};
// Enum for module arguments
@@ -178,6 +198,8 @@ typedef enum _ARG_ID {
ARG_IMAGEQUALITY,
+ ARG_MODEL_NAME,
+
ARG_UNKNOWN
} ARG_ID;
@@ -259,6 +281,9 @@ typedef enum _ATTR_ID {
ATTR_PERF_LEVEL,
ATTR_IMAGE_QUALITY,
+
+ ATTR_MODEL_NAME,
+
} ATTR_ID;
@@ -309,10 +334,35 @@ struct nvidia_s {
ATTR_ID attribute;
char *token;
SEARCH_ID search;
+// added new field for GPU id
+ int gpu_id;
+};
+
+//Cache by value
+struct nvidia_c_value {
+ int memtotal = -1;
+ int gputempthreshold = -1;
+};
+
+//Cache by string
+struct nvidia_c_string {
+ int nvclockmin = -1;
+ int nvclockmax = -1;
+ int memclockmin = -1;
+ int memclockmax = -1;
+ int memTransferRatemin = -1;
+ int memTransferRatemax = -1;
+ int perfmin = -1;
+ int perfmax = -1;
};
static Display *nvdisplay;
+// Maximum number of GPU connected:
+// For cache default value: choosed a model of direct access to array instead of list for speed improvement
+// value based on the incoming quad Naples tech having 256 PCIe lanes available
+const int MAXNUMGPU=64;
+
namespace {
class nvidia_display_setting: public conky::simple_config_setting {
@@ -365,12 +415,34 @@ int set_nvidia_query(struct text_object *obj, const char *arg, unsigned int spec
{
struct nvidia_s *nvs;
int aid;
+ int ilen;
+
+
+
// Initialize global struct
obj->data.opaque = malloc(sizeof(struct nvidia_s));
nvs = static_cast(obj->data.opaque);
memset(nvs, 0, sizeof(struct nvidia_s));
+ // Added new parameter parsing GPU_ID as 0,1,2,..
+ // if no GPU_ID parameter then default to 0
+ nvs->gpu_id = 0;
+ char *strbuf = strdup(arg);
+ char *p = strrchr(strbuf, ' ');
+ if (p && *(p + 1)) {
+ nvs->gpu_id = atoi(p+1);
+ if( (nvs->gpu_id > 0) || !strcmp(p+1,"0") ) {
+ ilen = strlen(strbuf);
+ ilen = ilen - strlen(p);
+ strbuf[ilen] = 0;
+ arg = strbuf;
+ }
+ }
+
+ // If the value is negative it is set to 0
+ if (nvs->gpu_id < 0) nvs->gpu_id = 0;
+
// Extract arguments for nvidiabar, etc, and run set_nvidia_query
switch (special_type) {
case BAR:
@@ -393,9 +465,11 @@ int set_nvidia_query(struct text_object *obj, const char *arg, unsigned int spec
if (strcmp(arg, translate_module_argument[aid]) == 0)
break;
}
- //fprintf(stderr, "parameter: %s -> aid: %d\n", arg, aid);
+
+ // free the string buffer after arg is not anymore needed
+ if (strbuf != NULL) free(strbuf);
- // Save pointers to the arg and command strings for dubugging and printing
+ // Save pointers to the arg and command strings for debugging and printing
nvs->arg = translate_module_argument[aid];
nvs->command = translate_nvidia_special_type[special_type];
@@ -582,6 +656,12 @@ int set_nvidia_query(struct text_object *obj, const char *arg, unsigned int spec
nvs->target = TARGET_SCREEN;
nvs->attribute = ATTR_IMAGE_QUALITY;
break;
+
+ case ARG_MODEL_NAME:
+ nvs->query = QUERY_STRING;
+ nvs->target = TARGET_GPU;
+ nvs->attribute = ATTR_MODEL_NAME;
+ break;
default: // Unknown/invalid argument
// Error printed by core.cc
@@ -590,7 +670,7 @@ int set_nvidia_query(struct text_object *obj, const char *arg, unsigned int spec
return 0;
}
-// Return the amount of targets present (or -1 on error)
+// Return the amount of targets present or raise error)
static inline int get_nvidia_target_count(Display *dpy, TARGET_ID tid)
{
int num_tgts;
@@ -598,48 +678,41 @@ static inline int get_nvidia_target_count(Display *dpy, TARGET_ID tid)
num_tgts = -1;
}
- return num_tgts;
-}
-
-// Exit if we are unable to get targets of type tid on display dpy
-void check_nvidia_target_count(Display *dpy, TARGET_ID tid, ATTR_ID aid)
-{
- int num_tgts = get_nvidia_target_count(dpy, tid);
-
if(num_tgts < 1) {
- // Print error and exit if there's not enough targets to query
+ // Print error and exit if there's no NVIDIA's GPU
CRIT_ERR(NULL, NULL, "%s:"
"\n Trying to query Nvidia target failed (using the propietary drivers)."
"\n Are you sure they are installed correctly and a Nvidia GPU is in use?"
- "\n (display: %d, target_id: %d, target_count: %d, attribute_id: %d)"
- , __func__, dpy, tid, num_tgts, aid);
+ "\n (display: %d,Nvidia target_count: %d)"
+ , __func__, dpy, num_tgts );
}
+
+ return num_tgts;
}
-static int cache_nvidia_value(TARGET_ID tid, ATTR_ID aid, Display *dpy, int *value)
+static int cache_nvidia_value(TARGET_ID tid, ATTR_ID aid, Display *dpy, int *value, int gid)
{
- static int memtotal = -1;
- static int gputempthreshold = -1;
+ static nvidia_c_value ac_value[MAXNUMGPU];
if (aid == ATTR_MEM_TOTAL) {
- if (memtotal < 0) {
- if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], 0, 0, translate_nvidia_attribute[aid], value)){
+ if (ac_value[gid].memtotal < 0) {
+ if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], gid, 0, translate_nvidia_attribute[aid], value)){
NORM_ERR("%s: Something went wrong running nvidia query (tid: %d, aid: %d)", __func__, tid, aid);
return -1;
}
- memtotal = *value;
+ ac_value[gid].memtotal = *value;
} else {
- *value = memtotal;
+ *value = ac_value[gid].memtotal;
}
} else if (aid == ATTR_GPU_TEMP_THRESHOLD) {
- if (gputempthreshold < 0) {
- if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], 0, 0, translate_nvidia_attribute[aid], value)){
+ if (ac_value[gid].gputempthreshold < 0) {
+ if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], gid, 0, translate_nvidia_attribute[aid], value)){
NORM_ERR("%s: Something went wrong running nvidia query (tid: %d, aid: %d)", __func__, tid, aid);
return -1;
}
- gputempthreshold = *value;
+ ac_value[gid].gputempthreshold = *value;
} else {
- *value = gputempthreshold;
+ *value = ac_value[gid].gputempthreshold;
}
}
@@ -647,22 +720,19 @@ static int cache_nvidia_value(TARGET_ID tid, ATTR_ID aid, Display *dpy, int *val
}
// Retrieve attribute value via nvidia interface
-static int get_nvidia_value(TARGET_ID tid, ATTR_ID aid)
+static int get_nvidia_value(TARGET_ID tid, ATTR_ID aid, int gid)
{
Display *dpy = nvdisplay ? nvdisplay : display;
int value;
- // Check for issues
- check_nvidia_target_count(dpy, tid, aid);
-
// Check if the aid is cacheable
if (aid == ATTR_MEM_TOTAL || aid == ATTR_GPU_TEMP_THRESHOLD) {
- if (cache_nvidia_value(tid, aid, dpy, &value)) {
+ if (cache_nvidia_value(tid, aid, dpy, &value, gid)) {
return -1;
}
// If not, then query it
} else {
- if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], 0, 0, translate_nvidia_attribute[aid], &value)){
+ if(!dpy || !XNVCTRLQueryTargetAttribute(dpy, translate_nvidia_target[tid], gid, 0, translate_nvidia_attribute[aid], &value)){
NORM_ERR("%s: Something went wrong running nvidia query (tid: %d, aid: %d)", __func__, tid, aid);
return -1;
}
@@ -680,84 +750,75 @@ static int get_nvidia_value(TARGET_ID tid, ATTR_ID aid)
// Retrieve attribute string via nvidia interface
-static char* get_nvidia_string(TARGET_ID tid, ATTR_ID aid)
+static char* get_nvidia_string(TARGET_ID tid, ATTR_ID aid, int gid)
{
Display *dpy = nvdisplay ? nvdisplay : display;
char *str;
- // Check for issues
- check_nvidia_target_count(dpy, tid, aid);
-
// Query nvidia interface
- if (!dpy || !XNVCTRLQueryTargetStringAttribute(dpy, translate_nvidia_target[tid], 0, 0, translate_nvidia_attribute[aid], &str)) {
- NORM_ERR("%s: Something went wrong running nvidia string query (tid: %d, aid: %d)", __func__, tid, aid);
+ if (!dpy || !XNVCTRLQueryTargetStringAttribute(dpy, translate_nvidia_target[tid], gid, 0, translate_nvidia_attribute[aid], &str)) {
+ NORM_ERR("%s: Something went wrong running nvidia string query (tid: %d, aid: %d, GPU %d)", __func__, tid, aid, gid);
return NULL;
}
- //fprintf(stderr, "%s", str);
+ //fprintf(stderr, "checking get_nvidia_string-> '%s'", str);
return str;
}
-static int cache_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEARCH_ID search, int *value, int update)
-{
- static int nvclockmin = -1;
- static int nvclockmax = -1;
- static int memclockmin = -1;
- static int memclockmax = -1;
- static int memTransferRatemin = -1;
- static int memTransferRatemax = -1;
- static int perfmin = -1;
- static int perfmax = -1;
+static int cache_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEARCH_ID search, int *value, int update, int gid)
+{
+ static nvidia_c_string ac_string[MAXNUMGPU];
if (update) {
- if (strcmp(token, (char*) "nvclockmin") == 0 && nvclockmin < 0){
- nvclockmin = *value;
- } else if (strcmp(token, (char*) "nvclockmax") == 0 && nvclockmax < 0){
- nvclockmax = *value;
- } else if (strcmp(token, (char*) "memclockmin") == 0 && memclockmin < 0){
- memclockmin = *value;
- } else if (strcmp(token, (char*) "memclockmax") == 0 && memclockmax < 0){
- memclockmax = *value;
- } else if (strcmp(token, (char*) "memTransferRatemin") == 0 && memTransferRatemin < 0){
- memTransferRatemin = *value;
- } else if (strcmp(token, (char*) "memTransferRatemax") == 0 && memTransferRatemax < 0){
- memTransferRatemax = *value;
+ if (strcmp(token, (char*) "nvclockmin") == 0 && ac_string[gid].nvclockmin < 0){
+ ac_string[gid].nvclockmin = *value;
+ } else if (strcmp(token, (char*) "nvclockmax") == 0 && ac_string[gid].nvclockmax < 0){
+ ac_string[gid].nvclockmax = *value;
+ } else if (strcmp(token, (char*) "memclockmin") == 0 && ac_string[gid].memclockmin < 0){
+ ac_string[gid].memclockmin = *value;
+ } else if (strcmp(token, (char*) "memclockmax") == 0 && ac_string[gid].memclockmax < 0){
+ ac_string[gid].memclockmax = *value;
+ } else if (strcmp(token, (char*) "memTransferRatemin") == 0 && ac_string[gid].memTransferRatemin < 0){
+ ac_string[gid].memTransferRatemin = *value;
+ } else if (strcmp(token, (char*) "memTransferRatemax") == 0 && ac_string[gid].memTransferRatemax < 0){
+ ac_string[gid].memTransferRatemax = *value;
- } else if (strcmp(token, (char*) "perf") == 0 && memTransferRatemax < 0){
+ } else if (strcmp(token, (char*) "perf") == 0 && ac_string[gid].memTransferRatemax < 0){
if (search == SEARCH_MIN) {
- perfmin = *value;
+ ac_string[gid].perfmin = *value;
} else if (search == SEARCH_MAX) {
- perfmax = *value;
+ ac_string[gid].perfmax = *value;
}
}
-
+
} else {
if (strcmp(token, (char*) "nvclockmin") == 0){
- *value = nvclockmin;
+ *value = ac_string[gid].nvclockmin;
} else if (strcmp(token, (char*) "nvclockmax") == 0){
- *value = nvclockmax;
+ *value = ac_string[gid].nvclockmax;
} else if (strcmp(token, (char*) "memclockmin") == 0){
- *value = memclockmin;
+ *value = ac_string[gid].memclockmin;
} else if (strcmp(token, (char*) "memclockmax") == 0){
- *value = memclockmax;
+ *value = ac_string[gid].memclockmax;
} else if (strcmp(token, (char*) "memTransferRatemin") == 0){
- *value = memTransferRatemin;
+ *value = ac_string[gid].memTransferRatemin;
} else if (strcmp(token, (char*) "memTransferRatemax") == 0){
- *value = memTransferRatemax;
+ *value = ac_string[gid].memTransferRatemax;
} else if (strcmp(token, (char*) "perf") == 0){
if (search == SEARCH_MIN) {
- *value = perfmin;
+ *value = ac_string[gid].perfmin;
} else if (search == SEARCH_MAX) {
- *value = perfmax;
+ *value = ac_string[gid].perfmax;
}
}
+
}
return 0;
}
// Retrieve token value from nvidia string
-static int get_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEARCH_ID search)
+static int get_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEARCH_ID search, int gid)
{
char *str;
char *kvp, *key, *val;
@@ -767,13 +828,13 @@ static int get_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEAR
value = -1;
// Checks if the value is cacheable and is already loaded
- cache_nvidia_string_value(tid, aid, token, search, &value, 0);
+ cache_nvidia_string_value(tid, aid, token, search, &value, 0, gid);
if ( value != -1 ) {
return value;
}
// Get string via nvidia interface
- str = get_nvidia_string(tid, aid);
+ str = get_nvidia_string(tid, aid, gid);
// Split string into 'key=value' substrings, split substring
// into key and value, from value, check if token was found,
@@ -805,7 +866,7 @@ static int get_nvidia_string_value(TARGET_ID tid, ATTR_ID aid, char *token, SEAR
}
// This call updated the cache for the cacheable values;
- cache_nvidia_string_value(tid, aid, token, search, &value, 1);
+ cache_nvidia_string_value(tid, aid, token, search, &value, 1, gid);
// TESTING - print raw string if token was not found;
// string has to be queried again due to strtok_r()
@@ -827,6 +888,11 @@ void print_nvidia_value(struct text_object *obj, char *p, int p_max_size)
struct nvidia_s *nvs = static_cast(obj->data.opaque);
int value, temp1, temp2;
char* str;
+
+ Display *dpy = nvdisplay ? nvdisplay : display;
+
+ //num_GPU calculated only once based on the physical target
+ static int num_GPU = get_nvidia_target_count(dpy, TARGET_GPU) - 1;
// Assume failure
value = -1;
@@ -834,20 +900,25 @@ void print_nvidia_value(struct text_object *obj, char *p, int p_max_size)
// Perform query
if (nvs != NULL) {
+
+ //Reduce overcommitted GPU number to last GPU
+ if( nvs->gpu_id > num_GPU ) nvs->gpu_id = num_GPU;
+
+ //Execute switch by query type
switch (nvs->query) {
case QUERY_VALUE:
- value = get_nvidia_value(nvs->target, nvs->attribute);
+ value = get_nvidia_value(nvs->target, nvs->attribute, nvs->gpu_id);
break;
case QUERY_STRING:
- str = get_nvidia_string(nvs->target, nvs->attribute);
+ str = get_nvidia_string(nvs->target, nvs->attribute, nvs->gpu_id);
break;
case QUERY_STRING_VALUE:
- value = get_nvidia_string_value(nvs->target, nvs->attribute, nvs->token, nvs->search);
+ value = get_nvidia_string_value(nvs->target, nvs->attribute, nvs->token, nvs->search, nvs->gpu_id);
break;
case QUERY_SPECIAL:
switch (nvs->attribute) {
case ATTR_PERF_MODE:
- temp1 = get_nvidia_value(nvs->target, nvs->attribute);
+ temp1 = get_nvidia_value(nvs->target, nvs->attribute, nvs->gpu_id);
switch (temp1) {
case NV_CTRL_GPU_POWER_MIZER_MODE_ADAPTIVE:
temp2 = asprintf(&str, "Adaptive");
@@ -866,20 +937,20 @@ void print_nvidia_value(struct text_object *obj, char *p, int p_max_size)
}
break;
case ATTR_MEM_FREE:
- temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED);
- temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL);
+ temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL, nvs->gpu_id);
value = temp2 - temp1;
break;
case ATTR_MEM_UTIL:
- temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED);
- temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL);
+ temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
}
break;
}
}
-
+
// Print result
if (value != -1) {
snprintf(p, p_max_size, "%d", value);
@@ -889,6 +960,7 @@ void print_nvidia_value(struct text_object *obj, char *p, int p_max_size)
} else {
snprintf(p, p_max_size, "N/A");
}
+
}
@@ -904,17 +976,17 @@ double get_nvidia_barval(struct text_object *obj) {
if (nvs != NULL) {
switch (nvs->attribute) {
case ATTR_UTILS_STRING: // one of the percentage utils (gpuutil, membwutil, videoutil and pcieutil)
- value = get_nvidia_string_value(nvs->target, ATTR_UTILS_STRING, nvs->token, nvs->search);
+ value = get_nvidia_string_value(nvs->target, ATTR_UTILS_STRING, nvs->token, nvs->search, nvs->gpu_id);
break;
case ATTR_MEM_UTIL: // memutil
case ATTR_MEM_USED:
- temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED);
- temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL);
+ temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_MEM_FREE: // memfree
- temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED);
- temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL);
+ temp1 = get_nvidia_value(nvs->target, ATTR_MEM_USED, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_MEM_TOTAL, nvs->gpu_id);
value = temp2 - temp1;
break;
case ATTR_FAN_SPEED: // fanspeed: Warn user we are using fanlevel
@@ -922,26 +994,26 @@ double get_nvidia_barval(struct text_object *obj) {
nvs->command, nvs->arg);
// No break, continue into fanlevel
case ATTR_FAN_LEVEL: // fanlevel
- value = get_nvidia_value(nvs->target, ATTR_FAN_LEVEL);
+ value = get_nvidia_value(nvs->target, ATTR_FAN_LEVEL, nvs->gpu_id);
break;
case ATTR_GPU_TEMP: // gputemp (calculate out of gputempthreshold)
- temp1 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP);
- temp2 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP_THRESHOLD);
+ temp1 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP_THRESHOLD, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_AMBIENT_TEMP: // ambienttemp (calculate out of gputempthreshold for consistency)
- temp1 = get_nvidia_value(nvs->target, ATTR_AMBIENT_TEMP);
- temp2 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP_THRESHOLD);
+ temp1 = get_nvidia_value(nvs->target, ATTR_AMBIENT_TEMP, nvs->gpu_id);
+ temp2 = get_nvidia_value(nvs->target, ATTR_GPU_TEMP_THRESHOLD, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_GPU_FREQ: // gpufreq (calculate out of gpufreqmax)
- temp1 = get_nvidia_value(nvs->target, ATTR_GPU_FREQ);
- temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "nvclockmax", SEARCH_MAX);
+ temp1 = get_nvidia_value(nvs->target, ATTR_GPU_FREQ, nvs->gpu_id);
+ temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "nvclockmax", SEARCH_MAX, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_MEM_FREQ: // memfreq (calculate out of memfreqmax)
- temp1 = get_nvidia_value(nvs->target, ATTR_MEM_FREQ);
- temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "memclockmax", SEARCH_MAX);
+ temp1 = get_nvidia_value(nvs->target, ATTR_MEM_FREQ, nvs->gpu_id);
+ temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "memclockmax", SEARCH_MAX, nvs->gpu_id);
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_FREQS_STRING: // mtrfreq (calculate out of memfreqmax)
@@ -951,13 +1023,13 @@ double get_nvidia_barval(struct text_object *obj) {
nvs->command, nvs->arg);
return 0;
}
- temp1 = get_nvidia_string_value(nvs->target, ATTR_FREQS_STRING, nvs->token, SEARCH_MAX);
- temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "memTransferRatemax", SEARCH_MAX);
+ temp1 = get_nvidia_string_value(nvs->target, ATTR_FREQS_STRING, nvs->token, SEARCH_MAX, nvs->gpu_id);
+ temp2 = get_nvidia_string_value(nvs->target, ATTR_PERFMODES_STRING, (char*) "memTransferRatemax", SEARCH_MAX, nvs->gpu_id);
if (temp2 > temp1) temp1 = temp2; // extra safe here
value = ((float)temp1 * 100 / (float)temp2) + 0.5;
break;
case ATTR_IMAGE_QUALITY: // imagequality
- value = get_nvidia_value(nvs->target, ATTR_IMAGE_QUALITY);
+ value = get_nvidia_value(nvs->target, ATTR_IMAGE_QUALITY, nvs->gpu_id);
break;
default: // Throw error if unsupported args are used