Commit 1f3e169b authored by Leonardo Solis's avatar Leonardo Solis
Browse files

#31, removed useless already-commented krnl attrib max_global_work_dim

parent dfd3066d
...@@ -3,9 +3,6 @@ ...@@ -3,9 +3,6 @@
// the genotype fed by any producer logic/kernel (IC, GG, LSs). // the genotype fed by any producer logic/kernel (IC, GG, LSs).
// Originally from: processligand.c // Originally from: processligand.c
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Conform( void Krnl_Conform(
__global const int* restrict KerConstStatic_rotlist_const, __global const int* restrict KerConstStatic_rotlist_const,
......
...@@ -217,9 +217,6 @@ float map_angle_360(float angle) ...@@ -217,9 +217,6 @@ float map_angle_360(float angle)
// Lamarckian Genetic-Algorithm (GA): GA + LS (Local Search) // Lamarckian Genetic-Algorithm (GA): GA + LS (Local Search)
// Originally from: searchoptimum.c // Originally from: searchoptimum.c
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_GA( void Krnl_GA(
__global const float* restrict GlobPopulationCurrentInitial, __global const float* restrict GlobPopulationCurrentInitial,
......
...@@ -15,9 +15,6 @@ ...@@ -15,9 +15,6 @@
// This has been later optimized, so now genotypes go directly // This has been later optimized, so now genotypes go directly
// from producer logic/kernel (IC, GG, LSs) to the consumer (Conform) kernel. // from producer logic/kernel (IC, GG, LSs) to the consumer (Conform) kernel.
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_IGL_Arbiter(/*unsigned char DockConst_num_of_genes*/ void Krnl_IGL_Arbiter(/*unsigned char DockConst_num_of_genes*/
#if !defined(SW_EMU) #if !defined(SW_EMU)
......
...@@ -15,9 +15,6 @@ ...@@ -15,9 +15,6 @@
// This has been later optimized, so now genotypes go directly // This has been later optimized, so now genotypes go directly
// from producer logic/kernel (IC, GG, LSs) to the consumer (Conform) kernel. // from producer logic/kernel (IC, GG, LSs) to the consumer (Conform) kernel.
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_IGL_Arbiter(/*unsigned char DockConst_num_of_genes*/ void Krnl_IGL_Arbiter(/*unsigned char DockConst_num_of_genes*/
#if !defined(SW_EMU) #if !defined(SW_EMU)
......
...@@ -6,9 +6,6 @@ ...@@ -6,9 +6,6 @@
// a very high value will be added to the current energy as a penalty. // a very high value will be added to the current energy as a penalty.
// Originally from: processligand.c // Originally from: processligand.c
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_InterE( void Krnl_InterE(
__global const float* restrict GlobFgrids, __global const float* restrict GlobFgrids,
......
...@@ -12,9 +12,6 @@ float sqrt_custom(const float x) ...@@ -12,9 +12,6 @@ float sqrt_custom(const float x)
// contributor-pairs. // contributor-pairs.
// Originally from: processligand.c // Originally from: processligand.c
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_IntraE( void Krnl_IntraE(
__constant float* restrict KerConstStatic_atom_charges_const, __constant float* restrict KerConstStatic_atom_charges_const,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS( void Krnl_LS(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS2( void Krnl_LS2(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS3( void Krnl_LS3(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS4( void Krnl_LS4(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS5( void Krnl_LS5(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS6( void Krnl_LS6(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS7( void Krnl_LS7(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS8( void Krnl_LS8(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_LS9( void Krnl_LS9(
unsigned short DockConst_max_num_of_iters, unsigned short DockConst_max_num_of_iters,
......
...@@ -2,9 +2,6 @@ ...@@ -2,9 +2,6 @@
// PRNG generators are implemented as Linear Feedback Shift Registers (LFSR) // PRNG generators are implemented as Linear Feedback Shift Registers (LFSR)
// All are 32-bit LFRS, feedback taps: 30, 20, 26, 25 // All are 32-bit LFRS, feedback taps: 30, 20, 26, 25
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_BT_ushort_float( void Krnl_Prng_BT_ushort_float(
unsigned int Host_seed1, unsigned int Host_seed1,
...@@ -77,9 +74,6 @@ void Krnl_Prng_BT_ushort_float( ...@@ -77,9 +74,6 @@ void Krnl_Prng_BT_ushort_float(
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_GG_uchar( void Krnl_Prng_GG_uchar(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -131,9 +125,6 @@ void Krnl_Prng_GG_uchar( ...@@ -131,9 +125,6 @@ void Krnl_Prng_GG_uchar(
} // while(valid != PIPE_STATUS_SUCCESS) } // while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_GG_float( void Krnl_Prng_GG_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -181,9 +172,6 @@ void Krnl_Prng_GG_float( ...@@ -181,9 +172,6 @@ void Krnl_Prng_GG_float(
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS123_ushort( void Krnl_Prng_LS123_ushort(
unsigned int Host_seed1, unsigned int Host_seed1,
...@@ -282,9 +270,6 @@ void Krnl_Prng_LS123_ushort( ...@@ -282,9 +270,6 @@ void Krnl_Prng_LS123_ushort(
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
// -------------------------------------------------------------------------- // --------------------------------------------------------------------------
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS_float( void Krnl_Prng_LS_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -330,9 +315,6 @@ void Krnl_Prng_LS_float( ...@@ -330,9 +315,6 @@ void Krnl_Prng_LS_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS2_float( void Krnl_Prng_LS2_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -378,9 +360,6 @@ void Krnl_Prng_LS2_float( ...@@ -378,9 +360,6 @@ void Krnl_Prng_LS2_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS3_float( void Krnl_Prng_LS3_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -426,9 +405,6 @@ void Krnl_Prng_LS3_float( ...@@ -426,9 +405,6 @@ void Krnl_Prng_LS3_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS4_float( void Krnl_Prng_LS4_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -474,9 +450,6 @@ void Krnl_Prng_LS4_float( ...@@ -474,9 +450,6 @@ void Krnl_Prng_LS4_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS5_float( void Krnl_Prng_LS5_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -522,9 +495,6 @@ void Krnl_Prng_LS5_float( ...@@ -522,9 +495,6 @@ void Krnl_Prng_LS5_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS6_float( void Krnl_Prng_LS6_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -570,9 +540,6 @@ void Krnl_Prng_LS6_float( ...@@ -570,9 +540,6 @@ void Krnl_Prng_LS6_float(
} // End of while(active) } // End of while(active)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS7_float( void Krnl_Prng_LS7_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -618,9 +585,6 @@ void Krnl_Prng_LS7_float( ...@@ -618,9 +585,6 @@ void Krnl_Prng_LS7_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS8_float( void Krnl_Prng_LS8_float(
unsigned int Host_seed, unsigned int Host_seed,
...@@ -666,9 +630,6 @@ void Krnl_Prng_LS8_float( ...@@ -666,9 +630,6 @@ void Krnl_Prng_LS8_float(
} // End of while(valid != PIPE_STATUS_SUCCESS) } // End of while(valid != PIPE_STATUS_SUCCESS)
} }
/*
__kernel __attribute__ ((max_global_work_dim(0)))
*/
__kernel __attribute__ ((reqd_work_group_size(1,1,1))) __kernel __attribute__ ((reqd_work_group_size(1,1,1)))
void Krnl_Prng_LS9_float( void Krnl_Prng_LS9_float(
unsigned int Host_seed, unsigned int Host_seed,
......
Supports Markdown
0% or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment