Mac requirement for Delphi XE4 Professional + Mobile Add-On - macos

I can't find information on what are the system requirements for Mac for the mobile add-on for Delphi XE4 (Professional).
My understanding is that XCode is still required to sign the app binary and to deploy it on a test device and/or App Store.
Our Mac is so old that the minimum requirement of OSX for the latest XCode won't even install.
I'd like to know if we need to have the Mac upgraded to accomodate this.

Ok - found the answer: https://www.embarcadero.com/products/rad-studio/rad-studio-data-sheet.pdf second page:
PC running Windows connected with an Intel-based Mac, with 2 GB RAM
or more, running OS X
10.8 (Mountain Lion) or 10.7 (Lion) over a local area network using a SSH, VNC or Windows file sharing solution
Apologies & much appreciated if you've attempted to help me:

XCode will still be required, so of course your Mac system will have to be updated to be able to support it.
The documentation for XE3 will give you the minimums requirements for that version, and you should require XE4 to at least require those minimum requirements. As it doesn't seem your existing system meets those, you can probably be sure it won't support the requirements for XE4.
You can also contact Embarcadero Sales with pre-sales question, with the email address you'd expect (Sales at the companyname.com).

Related

How to install XCode 9 GM on MacOS Sierra 10.12?

I actually built a Hackintosh to learn programming with Xcode. It runs on my Asus X555LA laptop. I downloaded the latest Xcode 9 GM build from the Apple Site (not from App store). After extracting, when I tried to install, it shows "You can't use this version of the application "Xcode" with this version of macOS; You have macOS 10.12. The application requires macOS 10.12.6 or later".
Is there any tweak to make it run on my Sierra 10.12 itself? I can't really think about upgrading the macOS version as it's a Hackintosh. I followed this guide to install macOS on my Asus laptop.
Xcode requires latest macOS, you have no choice, you need to upgrade the macOS version on your Hackintosh. Or better: Reinstall macOS in a recommended way on your PC, if you're doing Hackintosh... :)
The guide you linked is very poor... Never use premade install images, because these have been modified in an uncertain way, and you don't want to install a premade undocumented mess to your computer. It might be packed with threats, malwares, spy tools and so on.. It's the worst thing I can imagine in security aspect to install an OS image from uncertain source.
Also, there is no universal macOS installer for PCs - even though many are trying to find a way to create it: it's a bad idea and it will never succeed because there are so many PC parts, millions of differently built computers..
The only way to create a stable fully functional Hackintosh is to know your hardware and create an installer flash drive for that specific PC. First you have to download the latest macOS Sierra from AppStore, this is the only source that you can trust, because it's downloading from Apple's servers. Then install a small program, called Clover bootloader to the flash drive to make it bootable.
This is the only full and up to date guide for PC laptops. If you have questions, register to the linked site and start a new forum thread posting your questions. They will help you but please read this guide at least 3-4 times carefully because everything is described here.

To install Xcode in Windows XP

I would like to know if I can install Xcode with Windows XP.If possible please provide the document link also.Thank you very much for any help in advance...
It's certainly possible.
There are two routes;
Install OSx86 (aka iATKOS / Kalyway) on a second partition/disk and dual boot.
Run Mac OS X Server under VMWare.
The first route requires modifying (or using a pre-modified) image of Leopard that can be installed on a regular PC. This is not as hard as you would think, although your success/effort ratio will depend upon how closely the hardware in your PC matches that in Mac hardware - e.g. if you're running a Core 2 Duo on an Intel Motherboard, with a NVidia graphics card you are laughing. If you're running an AMD machine or something without SSE3 it gets a little more involved.
If you purchase (or already own) a version of Leopard then this is a gray area since the Leopard EULA states you may only run it on an "Apple Labeled" machine. As many point out if you stick an Apple sticker on your PC you're probably covered.
The second option is the more costly. The EULA for the workstation version of Leopard prevents it from being run under emulation and as a result there's no support in VMWare for this. Leopard server however CAN be run under emulation and can be used for desktop purposes. Leopard server and VMWare are expensive however.
If you're interested in option 1) I would suggest starting at Insanelymac and reading the OSx86 sections.
I do think you should consider whether the time you will invest is going to be worth the money you will save though. It was for me because I enjoy tinkering with this type of stuff and I started during the early iPhone betas, months before their App Store became available.
Alternatively you could pickup a low-spec Mac Mini from eBay. You don't need much horse power to run the SDK and you can always sell it on later if you decide to stop development or buy a better Mac.
No. You can not install XCode on a Windows machine. You need MacOS to run XCode.
Although you can install VMWare Server on your windows machine and then install MacOS on that virtual server and can install the XCode on that MacOS server.
But to install MacOS(VMWare server) and start working on it. your hardware must support virtualization.
To install VMWare following links might be useful.
http://www.petri.co.il/virtual_install_vmware_server.htm
http://www.virtuatopia.com/index.php/Installing_VMware_Server_2.0_on_Windows_Systems
Or there is always Google.
YOu can check if your CPU supports virtualization here.
Since Xcode is an software of Mac OS so its not possible to run Xcode without Mac OS. And for Mac OS you can dual boot your computer with a Mac OS or simply you can just use VMWare to install Mac OS in your laptop or PC. Just the necessary requirement for installing Mac OS through Dual boot or through VMWare is that you need really high configuration in your laptop or PC, Like atleast you require 5th Generation core processor with atleast 4 cores also you require more than 4GB RAM for better functioning of Mac OS. Mac OS can be installed with 2nd generation processors, Dual core, and 2 GB RAM but it will take too much time. Which is not worth installing Mac OS in your laptop or PC. And even after installing you won't be able to use Xcode efficiently with such a low configuration. So it is recommended to have 5th Generation processor with 4 core and atleast 4GB RAM.
If you have this configuration than you need some files and software to install Mac OS in your Laptop.
For installation through VMWare you can prefer to this video: https://www.youtube.com/watch?v=wodqGvug6e0
And I have the required file for the same as in video but not uploaded to the internet.

How to Virtualize iOS on a Windows 7 computer

I'm trying to learn to code in Objective C, but I am not ready to commit to purchasing a Mac for just this purpose, seeing as I have a good Dell laptop. My computer has VMware installed on it, but I do not know how to actually go through the steps, and I do not know what I need to install or purchase(if absolutely necessary). Eventually, I hope to get a Mac, so this won't be needed, but until then, I want to find a way to compile Objective C so I can actually learn how to use it. Thanks.
While it is technically possible, I'll tell you from experience:
To run Xcode and the iOS Simulator, you will need to have MacOS running. Oficially it only runs on Apple Hardware, buuut, you can always use virtualizers and Hackintosh. There are plenty of good tutorials online if you choose this method.
My advice: DON'T do it!
When I started to develop iOS apps, I had a Windows PC, but after trying lots of different approaches to virtualize MacOS, with none I had great results. I bought a MacBook on a sale, and it's just a better iOS development experience overall. It's priceless not to have to deal with driver searching and compatibility issues.
First virtualize MacOS X, inside then iOS with XCode. See http://www.macbreaker.com/2012/07/mountain-lion-virtualbox.html
Yes you can, you need to virtualize OSX inside you Windows, and then use XCode as the other stated, however it doesen't work with nearly all AMD CPU's but i think Dell uses mostly Intel. And a Mac costs alot. For example i don't have money for one, the cheapest one costs $1,600 where i come from, but if you are going to get a Mac, i would wait for that.
But they really are expensive, and i only need a Macbook for debugging.
[Assuming you already have the mac installation image or vmware file]
Sadly Vmware does not include support for mac in it's latest versions
So you'll have to patch(aka Vmware Unlocker for OS X) it.
Follow this link http://www.insanelymac.com/forum/files/file/20-vmware-unlocker-for-os-x/ (you'll be required to set up an account)
Or alternatively you can download the VirtualBox(https://www.virtualbox.org/)
Happy Coding
Use virtualbox instead of VMWare. If it does'nt work, install the extension pack for virtualbox.

confusion of how to make osx app backward compatible & how to test them

after reading the apple SDK guide
https://developer.apple.com/library/mac/#documentation/developertools/conceptual/cross_development/Overview/overview.html
I'm still confused of how to make the mac app backward compatible & how to test them properly
I have an app, I run it and tested it on Mountain Lion 10.8 without any problem, however I want to make this app backward compatible so that other users can run it on a mac 10.6 - 10.7 machine.
I have an apple developer id and I can download the old versions of 10.7 and 10.6, but the problem is, I have a 2011 macbook air which is currently running 10.8, and that's the only apple machine that I have. Can I test the 10.7 and 10.6 by using vmware or parallels?
in my project settings, I set the target deployment to 10.6 (as I want 10.6 users to run my app), but should I set my SDK to 10.8 or 10.7? if I set the SDK to 10.8 but having the target deployment set to 10.6, if I fix all the xcode warnings will it run successfully on 10.6??
from the SDK drop down, I can only set to 10.8 or 10.7, but 10.6 is missing, how do I fix that?
thanks in advance
I develop on a 10.8 box and support back to 10.5. Just a couple of months ago we dropped 10.4 PPC support, and I'm still cleaning out some of the 10.2-specific code. This may get a little rant-y, but I've been doing old versions for a long time. I have some opinions on the matter.
No matter what Apple says in their docs, if you want to support 10.6, then build with the 10.6 SDK. Do not rely on distribution target.
I have had this discussion with the Xcode engineers, and while they hold to Apple's party line that you should always build with the latest SDK, they also acknowledge that it's generally insane to do so. If you build against the 10.8 SDK and mark your deployment target at 10.6, you will get no warnings for using methods that do not exist on 10.6. The only way you will discover that you've used a nonexistent method is that it might give you strange bugs when run on 10.6. That's insane.
Remember, OS X doesn't crash when you send an unknown selector. It just aborts the current runloop. So the bugs are even harder to track down then on iOS, where it crashes the app.
Sure, you can do weak linking. Talk about dangerous.... Yes, there are a few times this is useful, but the compiler gives you no warning if you don't do it correctly. If I'm going to do weak linking like this, I go the other way, linking against the old SDK and copying the new function's prototype into my implementation. That way I have documentation of every function I think I'm going to weak-link.
Download the old SDKs and symlink them into your Xcode distribution.
Guard them jealously. Apple will try to delete them every time you upgrade Xcode. Make your own copies and stick them in /SDKs or somewhere else away from Xcode. I provide a script called fix-xcode to manage the symlinks automatically. Am I bitter at Apple for their relentless insistance on deleting my old SDKs? Yes, I am.
You can run 10.6 Server in a VM legally. You can run 10.7+ Desktop in a VM legally. These are good ways to test your code.
Or you can do what I do and have a small pile of old MacBooks each with two or three partitions on them that you reboot all the time.
Now that 10.7 comes from App Store, it's a little harder to make VMs. My strong recommendation is to snapshot your image immediately after install, and make a clean backup copy of it. You'll want to be able to clone that image from time to time when you need to get back to a "raw" machine.
Get in the habit of squirreling away SDKs as they come out. 10.8 will be old some day. You might as well make a copy now while it's easy.
Whether you support individual dot-releases or not, it can be very helpful to keep around the upgrade packages for individual dot releases. When you encounter customers running non-current releases, it's nice to be able to check whether an "unreproducible" bug in fact is easily reproducible on their specific version. Whether this is worth it or not depends heavily on your product and customers. It was a life-saver for me when 10.4.11 made major changes to WebKit during a dot release...
Invest in a small NAS or a big external USB drive (though I've had trouble with those failing when used extensively, so I prefer a RAID). You'll need the space. You want to hold onto lots of VMs and lots of SDKs and sometimes even old versions of Xcode.
Adding to Rob Napier's great in-depth answer:
To use an old SDK, put the SDK (or a symlink) to it here:
/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Developer/SDKs
With XCode 7.3 or later, you need you to open this file and change "MinimumSDKVersion" (otherwise XCode will refuse to use the old SDK):
/Applications/Xcode.app/Contents/Developer/Platforms/MacOSX.platform/Info.plist
You can install multiple versions of Mac OS on a single machine, booting between each.
The SDK should be the latest (10.8).
See 2.
One alternative to 1 that I've considered (I am in the same boat) is to create a Snow Leopard Hackintosh using an old PC and just installing Lion and Mountain Lion on my MBP.
You need to do these settings :
1.Set the Base SDK to Current version of Mac (ex. 10.7)
2.Set the Deployment SDK to older version (ex.1.4)

MonoTouch on OS X VM within Windows?

We're getting into iOS development with MonoTouch. All of our machines are Mac Pros with Windows 7 installed via BootCamp. I'm not crazy about rebooting into OS X just to access the MonoTouch IDE. I'm wondering if it's legal and possible to install OS X on a VM within Windows (if I'm already on Apple hardware, it should be ok, right?). Any other issues with Apple's SDK in a VM (I heard they do some hardware checking of some sort). Thanks in advance for any suggestions!
You can't really run OSX on a VM under windows without going the hacking route. The only way to properly virtualize OSX is to run OSX Server under OSX itself, which is not what you want.
The best option for you is to do what I do: run OSX on your Mac, then use something like VMWare or Parallels to run the Windows you have on your BootCamp as a VM. Works beautifully.
Yup, Eduardo is right, running OSX under non-apple hardware is considered illegal according to apple's license. Moreover, you may run into some issues when creating your developer's account or sumbitting apps.
However, if you still want go the hack way, you can refer to osx86project or just search google for "how to create a hackintosh".

Resources