Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 5 years ago.
Improve this question
i found an interesting effect on this site: https://iohk.io/
If you move your mouse around the effect will follow and change its colors, im really fascinated and wonder if there is a term for this kind of effect i can look into so i can learn how this was done.
While checking the source code i found the term "canvas", but its a huge topic.
Im not a developer myself and only know HTML/CSS right now, so please keep this in mind when answering.
Thanks in Advance.
http://vincentgarreau.com/particles.js/
That is the site from the creator, i haven't look in to it that much but all the information you need should be there.
You can use the menu to the right to change the way it looks.
Having looked at the page I initially thought of WebGL. On inspecting the page my suspicions were confirmed.
The page you linked to is using a ParticlesJS which looks like a library over WebGL. Given you are not a developer then this may be a little beyond you.
Related
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 2 years ago.
Improve this question
are there difference between unclickable or clickable images?
for example page load time and another things because I see people use them
So I was wondering which one is better
unclickable images give you a better page load time, if you want a page to load fast make them unclickable since page load times really influence for example customers when buying stuff online.
please see :
https://smallbiztrends.com/2019/02/page-load-speed-impacts-buying-decisions.html
Closed. This question needs debugging details. It is not currently accepting answers.
Edit the question to include desired behavior, a specific problem or error, and the shortest code necessary to reproduce the problem. This will help others answer the question.
Closed 6 years ago.
Improve this question
I have emojis used in the different segments for my Segmented Control. When I select a segment, the selected segment highlights (indicates which one was selected), and the emoji suddenly becomes all white. Is there anyway to fix this issue? Thanks!
I use follow code and it worked for me.
OC:
[segmentedControl setTitleTextAttributes:#{NSForegroundColorAttributeName:[UIColor whiteColor]} forState:UIControlStateSelected];
Swift:
let segmentedControl = UISegmentedControl(items: ["π","π’","moon"])
segmentedControl.setTitleTextAttributes([NSForegroundColorAttributeName:UIColor.whiteColor()], forState:.Selected)
I believe this is because the coding for emoji is not compatible with UIViews.
What you can do is select Momentary for the State. It will still disappear on click but will quickly come back.
Unfortunately the segment won't stay hilighted blue like you probably want though.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 8 years ago.
Improve this question
Does anyone know which API that Windows uses to draw the window title in Win 7(Aero theme) and Win 8?
I've tried to check with API Monitor tool, but don't see any API to draw the title.
Please advise.
Thanks.
It's been a long time since I did these things, but just googling it (three search phrase attempts) I ended up at DrawFrameControl.
Disclaimer: haven't tested.
Update: as noted in the comments DrawFrameControl won't do the job. And simple testing indicates that neither will DrawCaption. Apparently one has to use the theme functionlity, for which I found an example of drawing a window caption at (http://msdn.microsoft.com/en-us/library/windows/desktop/bb688195%28v=vs.85%29.aspx#appendixb). Essentially this code uses GetThemeSysFont to obtain the caption font and then DrawThemeTextEx to draw the caption.
New disclaimer: haven't tested the GetThemeSysFont plus DrawThemeTextEx combo, since this turned into quite a bit of work!
Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 9 years ago.
Improve this question
I have a set of images on a data set and I want to create an application(on android initially) so that the application would be able to take an image and match it with an image on my data set. I tried using image recognition algorithms but I had very poor recognition chances. Is there any way to unobtrusively embed data on the image itself such that when a printed version of the image is scanned by the application, it would be able to extract the embedded data? I need some direction in this regard because I seem to be punching in the dark since I haven't worked on anything like this before.
Thanks
You can try with OpenCV4Android, here is the introductory tutorial. Following SO will give you some more- Looking for OpenCV android tutorial
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 9 years ago.
Improve this question
How does Udacity.com (new learning website) create presentation where the text is in front the video is in back? Is there any specific software they use to prepare the presentations?
Udacity is using Wacom Cintiq tablet for producing those presentations.
Some extra information to add on Suhair's info about the Wacom Cintiq tablet being used.
Katy Reichelt is in charge of the video editing process and she explains how it has evolved and how it is done in her blog.
In that post you will find a link to professor's John Regehr recollection of how the class is recorded:
The recording process is this. You shut yourself into a dark, soundproofed room, put on a mic, and turn on a video camera and a program that records the tablet. Then, record written and spoken material. The editors can remove bad takes and pauses, which after years of lecturing to a live audience was hard for me to internalize. However, in the end it was very freeing since I could draw pictures, change colors, lookup details on Wikipedia, and attempt to write neatly (always a struggle for me) β all without boring anyone except the editors.The editors are 100% crucial to what Udacity is doing, and the difference between unedited material and the final output is striking. I canβt say enough about how good their work is.
My best guess is something like the following:
Link