llvm | Library for interacting with LLVM IR in pure Go | Compiler library
kandi X-RAY | llvm Summary
kandi X-RAY | llvm Summary
Library for interacting with LLVM IR in pure Go.
Support
Quality
Security
License
Reuse
Top functions reviewed by kandi - BETA
Currently covering the most popular Java, JavaScript and Python libraries. See a Sample of llvm
llvm Key Features
llvm Examples and Code Snippets
Community Discussions
Trending Discussions on llvm
QUESTION
ANSWER
Answered 2022-Apr-08 at 15:31Consider a typical use case of a std::any
: You pass it around in your code, move it dozens of times, store it in a data structure and fetch it again later. In particular, you'll likely return it from functions a lot.
As it is now, the pointer to the single "do everything" function is stored right next to the data in the any
. Given that it's a fairly small type (16 bytes on GCC x86-64), any
fits into a pair of registers. Now, if you return an any
from a function, the pointer to the "do everything" function of the any
is already in a register or on the stack! You can just jump directly to it without having to fetch anything from memory. Most likely, you didn't even have to touch memory at all: You know what type is in the any
at the point you construct it, so the function pointer value is just a constant that's loaded into the appropriate register. Later, you use the value of that register as your jump target. This means there's no chance for misprediction of the jump because there is nothing to predict, the value is right there for the CPU to consume.
In other words: The reason that you get the jump target for free with this implementation is that the CPU must have already touched the any
in some way to obtain it in the first place, meaning that it already knows the jump target and can jump to it with no additional delay.
That means there really is no indirection to speak of with the current implementation if the any
is already "hot", which it will be most of the time, especially if it's used as a return value.
On the other hand, if you use a table of function pointers somewhere in a read-only section (and let the any
instance point to that instead), you'll have to go to memory (or cache) every single time you want to move or access it. The size of an any
is still 16 bytes in this case but fetching values from memory is much, much slower than accessing a value in a register, especially if it's not in a cache. In a lot of cases, moving an any
is as simple as copying its 16 bytes from one location to another, followed by zeroing out the original instance. This is pretty much free on any modern CPU. However, if you go the pointer table route, you'll have to fetch from memory every time, wait for the reads to complete, and then do the indirect call. Now consider that you'll often have to do a sequence of calls on the any
(i.e. move, then destruct) and this will quickly add up. The problem is that you don't just get the address of the function you want to jump to for free every time you touch the any
, the CPU has to fetch it explicitly. Indirect jumps to a value read from memory are quite expensive since the CPU can only retire the jump operation once the entire memory operation has finished. That doesn't just include fetching a value (which is potentially quite fast because of caches) but also address generation, store forwarding buffer lookup, TLB lookup, access validation, and potentially even page table walks. So even if the jump address is computed quickly, the jump won't retire for quite a long while. In general, "indirect-jump-to-address-from-memory" operations are among the worst things that can happen to a CPU's pipeline.
TL;DR: As it is now, returning an any
doesn't stall the CPU's pipeline (the jump target is already available in a register so the jump can retire pretty much immediately). With a table-based solution, returning an any
will stall the pipeline twice: Once to fetch the address of the move function, then another time to fetch the destructor. This delays retirement of the jump quite a bit since it'll have to wait not only for the memory value but also for the TLB and access permission checks.
Code memory accesses, on the other hand, aren't affected by this since the code is kept in microcode form anyway (in the µOp cache). Fetching and executing a few conditional branches in that switch statement is therefore quite fast (and even more so when the branch predictor gets things right, which it almost always does).
QUESTION
My flutter app run well, but when I try to upload the app to App Store by archive it:
Xcode -> Product -> Archive
it failed and get two errors
First one in flutter_inappwebview with following error message:
ANSWER
Answered 2022-Mar-22 at 07:22Downgrading Xcode from 13.3 to 13.2.1 solved my problems.
QUESTION
I have created an app using React Native and am trying to create an iOS app store build through Expo's eas-cli.
When running eas build --platform ios
the Fastlane build failed with unknown error
After checking the "Run Fastlane" section in the Expo build log, multiple errors are shown:
Error 1:
...ANSWER
Answered 2021-Oct-06 at 06:11There are a number of things to look into.
If you are running Expo in the SDK then no need for cocoa pods just the most up-to-date version of the CLI tool.
Run expo --version
to determine what version you are currently working with. Update if needed.
Adding a profile might be useful too. Along with checking your config. Configuring EAS Build with eas.json
eas build --platform ios --profile distribution
Also, be sure that all the apple certificates are active and connected to your Expo account for that project.
QUESTION
My app archive perfectly two weeks age (was the latest update I made).
After some minor changes in the flutter code (I did not add any packages or something), I am trying to archive again and I am getting some estrange errors and I have no clue where are coming from (those are from libraries, so no real changes on it). I am able to run it on the emulator and also I am able to build without any problem. I just can not archive.
This is the error:
...ANSWER
Answered 2022-Mar-20 at 09:40This is an issue that occurs in version 13.3 of Xcode. In Xcode 13.3, if you have a code that uses UI_USER_INTERFACE_IDIOM(), you will get an "Out of Memory" error when you run Archive. Changing "UI_USER_INTERFACE_IDIOM()" to "UIDevice.current.userInterfaceIdiom" resolves the error.
Currently, we have the following solutions.
- Modify the code,
- Downgrade to Xcode 13.2.1
- Wait for Apple to modify Xcode
References
QUESTION
A standard idiom is
...ANSWER
Answered 2022-Mar-14 at 22:53The boolean conversion operator for std::basic_istream
is explicit
. This means that instances of the type will not implicitly become a bool
but can be converted to one explicitly, for instance by typing bool(infile)
.
Explicit boolean conversion operators are considered for conditional statements, i.e. the expression parts of if
, while
etc. More info about contextual conversions here.
However, a return statement will not consider the explicit
conversion operators or constructors. So you have to explicitly convert that to a boolean for a return
.
QUESTION
I'm getting this error:
...ANSWER
Answered 2022-Feb-06 at 12:57Please, consider trying to follow the instruction from quarkus/faq.adoc at main · quarkusio/quarkus:
1. Native compilationNative executable fails on macOS with
error: unknown type name 'uint8_t'
Your macOS has the wrong
*.h
files compared to the OS and no gcc compilation will work. This can happen when you migrate from versions of the OS. See Cannot compile any C++ programs; error: unknown type name 'uint8_t'The solution is to
sudo mv /usr/local/include /usr/local/include.old
- Reinstall XCode for good measure
- (optional?)
brew install llvm
- generally reinstall your brew dependencies with native compilation
The executable should work now.
The purpose of (the idea behind) renaming the include
directory (from include
to include.old
) seems to be explained in the answer.
- Related question. May contain some possible solutions. Broken c++ std libraries on macOS High Sierra 10.13 - Stack Overflow.
QUESTION
Although High Sierra is no longer supported by Homebrew, but I need to install llvm@13
formula as a dependency for other formulas. So I tried to install it this way:
ANSWER
Answered 2021-Nov-26 at 08:27Install llvm with debug mode enabled:
QUESTION
I thought hot-patching assumed that overwriting any instruction that is 2 or more bytes long with a 2 byte jump is safe for concurrent execution of the same code.
So instruction fetch is assumed to be atomic.
Is it indeed atomic, taking into account that with prefixes it is possible to have more than 8 bytes instruction, and it can cross any aligned boundary? (Or does hot-patching rely on 16-byte alignment of function start? If so, what's with size over 8 bytes, anyway?)
The context: LLVM has interception of API functions in https://github.com/llvm/llvm-project/blob/main/compiler-rt/lib/interception/interception_win.cpp. This is used at least for Address Sanitizer, maybe for something else too. It implements HotPatch as 3rd method (line 61):
...ANSWER
Answered 2021-Dec-05 at 00:12Instruction fetch is not architecturally guaranteed to be atomic. Although, in practice, an instruction cache fill transaction is, by definition atomic, meaning that the line being filled in the cache cannot change before the transaction completes (which happens when the whole line is stored in the IFU, but not necessarily in the instruction cache itself). The instruction bytes are also delivered to the input buffer of the instruction predecode unit at some atomic granularity. On modern Intel processors, the instruction cache line size is 64 bytes and the input width of the predcode unit is 16 bytes with an address aligned on a 16-byte boundary. (Note that the 16 bytes input can be delivered to the predecode unit before the entire transaction of fetching the cache line containing these 16 bytes completes.) Therefore, an instruction aligned on a 16-byte boundary is guaranteed to be fetched atomically, together with at least one byte of the following contiguous instruction, depending on the size of the instruction. But this is a microarchitectural guarantee, not architectural.
It seems to me that by instruction fetch atomicity you're referring to atomicity at the granularity of individual instructions, rather than some fixed number of bytes. Either way, instruction fetch atomicity is not required for hotpatching to work correctly. It's actually impractical because instruction boundaries are not known at the time of fetch.
If instruction fetch is atomic, it may still be possible to fetch, execute, and retire the instruction being modified with only one of the two bytes being written (or none of the bytes or both of the bytes). The allowed orders in which writes reach GO depend on the effective memory types of the target memory locations. So hotpatching would still not be safe.
Intel specifies in Section 8.1.3 of the SDM V3 how self-modifying code (SMC) and cross-modifying code (XMC) should work to guarantee correctness on all Intel processors. Regarding SMC, it says the following:
To write self-modifying code and ensure that it is compliant with current and future versions of the IA-32 architectures, use one of the following coding options:
(* OPTION 1 *)
Store modified code (as data) into code segment;
Jump to new code or an intermediate location;
Execute new code;(* OPTION 2 *)
Store modified code (as data) into code segment;
Execute a serializing instruction; (* For example, CPUID instruction *)
Execute new code;The use of one of these options is not required for programs intended to run on the Pentium or Intel486 processors, but are recommended to ensure compatibility with the P6 and more recent processor families.
Note that the last statement is incorrect. The writer probably intended to say instead: "The use of one of these options is not required for programs intended to run on the Pentium or later processors, but are recommended to ensure compatibility with the Intel486 processors." This is explained in Section 11.6, from which I want to quote an important statement:
A write to a memory location in a code segment that is currently cached in the processor causes the associated cache line (or lines) to be invalidated. This check is based on the physical address of the instruction. In addition, the P6 family and Pentium processors check whether a write to a code segment may modify an instruction that has been prefetched for execution. If the write affects a prefetched instruction, the prefetch queue is invalidated. This latter check is based on the linear address of the instruction
Briefly, prefetch buffers are used to maintain instruction fetch requests and their results. Starting with the P6, they were replaced with streaming buffers, which have a different design. The manual still uses the term "prefetch buffers" for all processors. The important point here is that, with respect to what is guaranteed architecturally, the check in the prefetch buffers is done using linear addresses, not physical addresses. That said, probably all Intel processors do these checks using physical addresses, which can be proved experimentally. Otherwise, this can break the fundamental sequential program order guarantee. Consider the following sequence of operations being executed on the same processor:
QUESTION
Trying to install Python 3.10.0 on MacOS 11.6 (Intel) with pyenv 2.1.0 (from homebrew) fails with:
...ANSWER
Answered 2021-Oct-06 at 05:56Try env ARCHFLAGS="-arch x86_64"
(that works for me for other Python-related tools, I don't use pyenv myself).
Rationale: modern macOS supports 2 architectures: intel (x86_64) and m1 (arm-something). Compiling for only one architecture is easier.
QUESTION
I recently had the following question
regarding the function mean!
of the Statistics.jl
package.
The following bug is reported regarding the behavior of mean!
. As indicated, the mean!
function does not properly consider that its arguments may alias each other. In some such cases the result from mean!
is not correct:
ANSWER
Answered 2021-Nov-14 at 13:05mean!
behaves the way you observe because it internally calls sum!
.
Now sum!
behaves this way for the following reason. It was designed to perform summation without making any allocations. Therefore the first thing sum!
does is initializing the target vector to 0
(a neutral element of summation). After this is done your a
vector contains only 0
s, and thus later you get all 0
s also.
However, indeed it would make sense that the sum!
(and similar) functions docstring should mention that the target should not alias with the source. Here is another example of the same you have observed:
Community Discussions, Code Snippets contain sources that include Stack Exchange Network
Vulnerabilities
No vulnerabilities reported
Install llvm
Support
Reuse Trending Solutions
Find, review, and download reusable Libraries, Code Snippets, Cloud APIs from over 650 million Knowledge Items
Find more librariesStay Updated
Subscribe to our newsletter for trending solutions and developer bootcamps
Share this Page