Want this question answered?
usually computers uses 0s and 1s as the means of their instructions. may be in some computers 01010 indicates addition, 01011 indicates subtraction and vice versa.ultimately when we write our program in high level languages such as c or c++ they have to be converted into 0s and 1s or machine language to be executed by the system.
You, as a programmer, can use a string with 1s and and 0s (or any other content) in each and every programming language.
Generally speaking, it doesn't relate at all. If it did, it wouldn't be high-level, it would be machine-dependent. The relationship between the high-level code and the machine-code is ultimately determined by the language translator (compiler and/or interpreter) but, unlike assembly language which maps 1:1 with the microprocessor instruction set, compilers and interpreters are code generators and there is seldom a 1:1 relationship between the high-level source code and the machine code.
HLL is what we humans use. Its full name is High Level Language. MLL is what machines use. Its full name is Machine Level Language. The difference is, HLL is our language (e.g English) and MLL is machines language (binary code) Binary code is 0s and 1s so "A" could be 010 in MLL. MLL is understood only by the computer and not humans
decode
Actually computer does not understand any high level language. it understands only 1s and 0s i.e what we call machine language. So the high level language needs to be translate into machine language which is done by the compiler It doesn't work in itself, it has to be translated to machine code, that's what you call compilation.
Sometimes it means different things. People use machine code for sequence of 0s and 1s, and assembler is usually called machine languages. But for the most part people use these two phrases interchangeably.
usually computers uses 0s and 1s as the means of their instructions. may be in some computers 01010 indicates addition, 01011 indicates subtraction and vice versa.ultimately when we write our program in high level languages such as c or c++ they have to be converted into 0s and 1s or machine language to be executed by the system.
The language of 0s and 1s is called binary which is internally used by the computer system for performing different activities. The other levels of languages such as high level languages, assembly language are internally converted into binary language for the processing by the computers.
You, as a programmer, can use a string with 1s and and 0s (or any other content) in each and every programming language.
Yes.
Generally speaking, it doesn't relate at all. If it did, it wouldn't be high-level, it would be machine-dependent. The relationship between the high-level code and the machine-code is ultimately determined by the language translator (compiler and/or interpreter) but, unlike assembly language which maps 1:1 with the microprocessor instruction set, compilers and interpreters are code generators and there is seldom a 1:1 relationship between the high-level source code and the machine code.
Binary
HLL is what we humans use. Its full name is High Level Language. MLL is what machines use. Its full name is Machine Level Language. The difference is, HLL is our language (e.g English) and MLL is machines language (binary code) Binary code is 0s and 1s so "A" could be 010 in MLL. MLL is understood only by the computer and not humans
decode
davros
You must learn it. You can search for keyword : Assembly.