feat: add hlsl_functionality_1 to c interface options
This commit is contained in:
parent
e9f840ca11
commit
da82563c92
2 changed files with 6 additions and 0 deletions
|
|
@ -345,6 +345,11 @@ GLSLANG_EXPORT glslang_shader_t* glslang_shader_create(const glslang_input_t* in
|
|||
shader->shader->setEnvTarget(c_shader_target_language(input->target_language),
|
||||
c_shader_target_language_version(input->target_language_version));
|
||||
|
||||
#ifdef ENABLE_HLSL
|
||||
if (input->hlsl_functionality_1)
|
||||
shader->shader->setEnvTargetHlslFunctionality1();
|
||||
#endif
|
||||
|
||||
if (input->entrypoint != nullptr)
|
||||
shader->shader->setEntryPoint(input->entrypoint);
|
||||
|
||||
|
|
|
|||
|
|
@ -220,6 +220,7 @@ typedef struct glslang_input_s {
|
|||
const char* entrypoint; // This is what actually gets called by the GPU. Best to leave it at 'main' or something so opengl doesn't trip over itself.
|
||||
const char* source_entrypoint; // This just renames the source entrypoint in the code to 'entrypoint' and should be what is used to set different entrypoint names.
|
||||
int invert_y;
|
||||
int hlsl_functionality_1;
|
||||
int default_version;
|
||||
glslang_profile_t default_profile;
|
||||
int force_default_version_and_profile;
|
||||
|
|
|
|||
Loading…
Add table
Add a link
Reference in a new issue