Grab the config files I posted here
There are two - one for your server, and one to run on your phone. The way it works is you have the ffserver running on a server box with fast-enough CPU and network connectivity. You can send it any stream using any codecs that ffmpeg supports, and ffmpeg will transcode it to a dumbed-down H.264 that can play smoothly over a relatively slow network. This transcode can run at realtime or faster, it all depends on your server and what the input to ffmpeg is. This ffserver will make the data available over HTTP in FLV format.
Install the second ffserver.conf file on your phone, along with the ffserver binary I posted in this thread. You of course need to tailor the file to replace "my.server:8080" with the address and port that you're using for the ffserver on your server box. This ffserver on the phone is configured to grab the FLV over HTTP from your main server, and then re-stream over RTSP/RTP/UDP. You can play the video from your main server by opening rtsp://127.0.0.1/test.mp4 - the Meridian player makes this easy since it has an Open URL menu for this. The SeeJo player also lets you play URLs.
I don't think I can explain it any more simply than this. If you don't understand how to run ffmpeg, there's plenty of other docs on that.
If you specifically want to stream h,u,l,u videos, you'll need to grab the perl script I posted above and edit the ffmpeg command at the end of the script. The script I attached here was set up to publish a stream to the Darwin Streaming Server over rtsp. A simpler command is used to stream to ffserver.
When publishing to DSS, you have to provide all of the audio/video codec settings on the ffmpeg command line, and the stream gets pushed to rtsp://wherever/the/DSS/server/is/listening.
When publishing to ffserver, all of the audio/video codec settings are in the ffserver.conf file, so you only need to tell ffmpeg where to find its input, and send its output to http://wherever/the/ffserver/is/listening