Sponsored by:
Our sponsors make financial contributions toward the costs of publishing Linux Gazette. If you would like to become a sponsor of LG, e-mail us at sponsor@ssc.com.
The Whole Damn Thing 1 (text)
The Whole Damn Thing 2 (HTML)
are files containing the entire issue: one in text format, one in HTML.
They are provided
strictly as a way to save the contents as one file for later printing in
the format of your choice;
there is no guarantee of working links in the HTML version.
Got any great ideas for improvements! Send your comments, criticisms, suggestions and ideas.
This page written and maintained by the Editor of Linux Gazette, gazette@ssc.com
The Mailbag!Write the Gazette at gazette@ssc.com |
Contents: |
The last couple of months have been rather light on articles. It's been helpful to have the new chapters for Linux Installation and Getting Starting to include. So all you budding authors and Linux users out there, send me your stuff. Don't depend on our regular authors to fill the gap. We want to hear about all the neat tips and tricks you've found, as well as all the neat applications your are writing or working with. --Editor
Date: Sat Apr 19 07:29:14 1997
Subject: Searching for Information On Newsgroups
From: Roman, Roman@pussycat.ping.de
Hi folks!
I'm installing a very small news- and email system at my local university (peolpe there are studying arts, so there's no one to help me with this). I set up one computer with Linux 2.0.29 which is permanently connected to the 'internet' via ethernet. Then I want to connect a second PC which is installed in the hallway via nullmodem-cable for all the students to write and receive eMail. But the problem now is, that the provider (another part of the university) doesn't give us access to the newsgroups, so I want to set up at least some local newsgroups on this Linux-station.
But I just can't seem to find any documentation explaining how to set up local newsgroups. smtpd and nntpd are running, but the manpages won't tell anything about how to set up ng's (forgive me if I'm just too blind or stupid to find the obvious source of information).
So I don't want to bother you explaining me how to accomplish this task, but perhaps someone can at least tell me where to find the desired information.
Best regards, Roman.
Date: Thu Apr 24 11:44:40 1997
Subject: VGA_16 Server
From: Javier Viscain
Congratulations for the aim and contents of the Gazette. Here is an issue I've never seen addressed: the VGA_16 server maintains two monitors (the second monochrome with an Hercules card) but what only works is the mouse movement, which moves out of left and right to the other monitor, and console switching. No window on the monochrome gets focus. Things that moreless appear on the monochrome but don't work:
I think that the hardware absolute addressing is the normal VGA one (0A0000 to 0AFFFF) and 64K for the Hercules (0B0000 to 0BFFFF), which is correct. In adition, this server and the mono server are very buggy when with only the Hercules.
Any easy solution, or is it that this configuration has not been debugged?
TIA, Javier Vizcaino, Madrid, Spain.
Date: Sun 6 Apr 1997 11:54:42 -0400
Subject: Initilation Files
From Karl Easterly bigtexan@mindspring.com
As an article Idea, I think an overview of the major boot scripts would =
be helpful. The overview could include an objective view of the =
locations, functions, and nifty "tips and tricks" or such. Also, links =
to how-toos for each script would future simplify the learning curve for =
new users.
Another idea would be to do a chronological installations and = customization series of articles. Granted, hardware diversity might be = a problem and could possibly be subverted by starting the series as = though a working installation of Linux has already been installed. It = would proceed as a rough idea like this.
These are just stabs at a scheme, the actual order would have to be = hammered out before the series started, but in general, would be helpful = to have a step by step issue oriented series of articles concerning the = setup and customization of any linux installation.
Date: Sat, 1 Mar 1997 15:39:10 -0500
Subject: Ideas for Beginners
From: stephen jarvis 106363.2642@compuserve.com
Hello
I am 'the' absolute beginner.I have had a copy of Linux Slackware and a copy of "Linux configuration and installation " by P Volkerding et al for about two weeks.Prior to this I had dabbled at dos and wondered(?) at windows.But when I heard about Linux in a magazine it occured to me that it might be fun to have a go.And indeed it has been.
The only problem I have had is with regard to the man pages. In general they are technical to a degree that while appropriate for those who can follow the argument from end to end,are pretty debilitating for the newbie like me.Indeed I don't always get to the end.
Perseverence will no doubt pay off and I have expanded my collection of books already,to take advantage of the possibility of learning something about programming on Linux.But then I have always had the kind of curiousity that,while not enough to kill the cat,is enough to keep me in the book shop.The point I think is that the man pages themselves are a bit of a barrier to the wider useage of Linux.
No doubt others would say the detail and technical clout of this source of information is needed for those who want to make serious use of Linux.But not everyone who wants to escape from the soporific influence of Microsoft is that demanding or that knowledgeable.I think someone needs to pitch things at the introductory level.In the realms of 'this will get it going'and 'try this out'.Merely a more chatty approach would help remove the shiney armour of incomprehensibilty some pages deploy.
If this sounds a little unfair to the many people who have compiled ,man pages it is most definitley not meant to be.There is a need for accurate and complete information especially as Linux is a cooperative venture and everyone needs to have a common root of information.The question is how can the benefits of Linux be made widely known to people outside the existing network.What will grab their attention and take the gleam off Windows 95?Something more open to a wider audience perhaps.
This does not have to be completely bland and overly simple just in the range of every day usage.An approach that does not assume that everyone reading knows the meaning of every term on the page.People need an introduction to the language of Linux in the way that you might learn French or English.Start with very basic things and build up in stages.Don't launch straight into 'How To Compile Your Kernal '.Ok thats important ,but I am sure most people still think a kernal is what you find inside a nut.I hope you are getting the general idea.
What us new people need is probably a collection of basic texts each about the length of a several page magazine article.Hopefully they would cover the things that a hardend Linux user would be embarrassed to ask about.'The kernal for beginners'.'Great now I can ask what it really is'.If this undertaking was started then I am sure that the end product of a few months could be published as a small book.Maybe you could publish it.I think there is a potential market.Many magazines recently covered the subject of Linux.That's how I got the bug.
Now it's true there are books already that cover Linux but there are not many on line man pages or magazine articles that give the beginner the feeling that they can actually get their system up and running easily.So if you really want to publish articles for absolute beginners bear in mind the kind of language that is used.
Regards Steve Jarvis
ps.. here's some ideas ' What is the kernal','The basic commands to get around bash','What are disk partitions and why bother',' To Umsdos or not. That's the question','Midnight Commander-an introduction','This is the easiest editor anybody ever used(insert your choice)','A glossary of general terms you'll find on a man page','These books are a good read(assorted titles)'.'How to get around an info text with less than 20 pages of instructions','Why the idea of a free and open o/s matters','X is not a horror film'.
Maybe these are a bit daft but they'd get my attention.They are the sort of things I'd like to know about.
Date: 04 Apr 97 19:02:21 EST
Subject: Technical Support
From: Dani Fricker 101550.3160@CompuServe.COM
first i wanna say thanx for the lj! great work and fun not even for linuxers! i need your help. for some reasons i have to identify a user on my webserver by his/her ip-address. fact is that users logon comes from different physical machines. that means that i have to assign something like a virtual ip-address to a users log name. something like a reversal masquerading. my ip-gateway connects my inner lan over two token ring network cards (sorry, not my idea!) with the internet (lan <-> tr0 <-> tr1 <-> internet). the masquerading forward roule of ipfwadm gives me the possibility to indicate a source and a destination address. do you see a possibility for an 'address assignment' between the two interfaces? if you do please let me know.
dani fricker
programmer
zurich-switzerland
Date: Mon, 07 Apr 1997 03:01:17 -0500
Subject:HELP with Man Pages
From: "Mauricio Naranjo N." davasgut@col2.telecom.com.co
Well, I have installed the linux toolkit / october 1996 and I have not been able to install the man pages for commonly used commands like cat, ls, and so on; instead I have installed the man pages for packages like, fvwm, midnight commander, ....
So, I installed man2.tgz, man3.tgz, manpgs.tgz, but I still have not been able to get installed the whole support for man; Can you tell me please, what's the matter???? Any kind of help would be great appreciated, and excuses for ignorance but I am new at this OS (finally I found a true one)
Mao
Date: Mon, 7 Apr 1997 15:43:21 -0400 (EDT)
Subject: Port Mouse
From: Jose, notDefined@novagate.com
Hi, Maybe you can help me with this. (I hope) I switched motherboards, from a zeos pentium 90 that used a serial mouse to a asus p/i-p55tvp4 motherboard that uses a port mouse. And now I can't get x-windows to run. Any ideas?
Date: Tue, 01 Apr 1997 04:26:04 -0600
Subject: Linux
From: Tred Riggs tred@oak.sfasu.edu
I am a college student attending Stephen F. Austin State University. I work in a Geographic Information Systems Laboratory (GIS) and we have been just using AIX machines. Howerver we do have a full blown linux pc and it is great. {Since then I stripped DOS off my PC and made me a full blown linux box, which works wonderful. We were considering to upgrade to all linux PC's in out lab because they were cheaper and faster than the AIX boxes, but we ran into a problem. The Software we need to run to make our GIS maps is not supported by ESRI, so we gave them a call. This is what they told us:
"Linux will not be a supported platform. They told me that product ports are user driven and there is not enough users wanting this OS."
I could not figure out how they could even say this when all you have to do is get on the web and see millions of people using linux. So here is what I want to happen. I need linux users to E-mail ESRI at buspartners@esri.com and tell them that you use linux and that there are many more people using linux too. ESRI needs to get there head out of Microsofts world and see what is going on in the real world.
Thanks for your time Linux Gazette,
Tred Riggs
Date: Thu, 3 Apr 97 22:40:23 BST
Subject:http://www.ssc.com/lg/index.html
From: Duncan Simpson dps@duncan.telstar.net
Given Micro$oft's tag line of "Yet another Web server powered by NT" maybe we should collect a list of people doing this sort of stuff on Linux. I can add 3 items myself http://mail.telstar.net is powered by Linux The telstar mail service described there is also powered by the same linux box Astra has switch from NT to Linux for its radius server. (NT was just too expensive and no better than Linux (Un*x)---the price diffrernece was *1000s* of pounds, each about 1.5 $ US). Both astra (and telstar.net) DNS servers are linux.
If the stats show that Linux is more popular for comercial web servers than NT, this would be something nice to be able to point out...
Duncan (-:
P.S. Any bets when Truetype fonts can be used for proper typesetiing. At present they lack litagures (fl and various other items that are tradionally rendered as single characters)?
P.P.S. The use of the present tense (switch) is apt because the change is happening now. (Despite a bug that is now not being exercised due to an attempt to eradicate it mail.telstar.net is more reliable than any of various NT machines at handling mail).
Date: Fri, 04 Apr 1997 15:57:12 -0600
Subject: Re: How to ftp Back Home
From: James Stansell james.stansell@wcom.com
The ifconfig command works, and may be the most authoritative on the subject (except I believe the PPP log also contains your current IP), but the ifconfig command returns a ton more information than I want.
So I ask my machine at work who I am:
who am i stansell ttyp6 Apr 4 15:51 (206.125.79.118)
I've inserted your example IP address where my actual address showed up. If the DNS at work does happen to know a name for my address, then it shows up instead of the IP.
--james
Date: Thu, 10 Apr 1997 17:08:38 -0500
Subject: Re:GV article
From: Larry Ayers ayers@vax2.rainis.net
To: Geoffrey Leach geoffrey@iname.com
Sorry the URL didn't work for you; I recently got an email
message from Helmut Geyer, the maintainer of the Debian GV
version and he included a URL for a new GV home-page:
http://wwwthep.physik.uni-mainz.de/~plass/gv/
The Debian version is in the /text section of the /i386 binary directory of any Debian mirror. Shouldn't be too hard to find.
Good luck!
Larry Ayers
From: Gary Masters gmasters@devcg.denver.co.us
I read your question in Linux Gazette regarding an X limitation to 8 bit color when the system has more than 14 megs of RAM. Where did you find that information? I ask because my system has 24 megs of RAM, and I run 16 bit color all the time. One difference between our systems is that I am using a Diamond Stealth 64 video card.
The place I tell X to run in 16 bit mode is in the file /usr/X11R6/bin/startx. There is a line in this file that begins with serverargs. I get 16 bit mode by giving "-bpp 16" as an argument in this line (e.g. serverargs="-bpp 16").
One problem I did have was that the OpenLook Window Manager (olwm) did not like 16 bpp mode. I solved this by switching to the OpenLook Virtual Window Manager (olvwm)[1]. I also had success using the Tab and FV Window Managers (twm & fvwm) in 16 bpp mode.
Coming from a SunOS background, I'm used to OpenLook.
Gary Masters
From: Gary Masters gmasters@devcg.denver.co.us
I read your question in the Linux Gazette regarding unwanted screen blanking under X after upgrading to a newer distribution of Linux. I had the same frustration. Apparently the X servers included in the Xfree86 version distributed with current Linux distributions has screen blanking compiled as a default behavior.
This behavior can be controlled with the -s option to the server. Look in the startx script for a line that begins with serverargs and add "-s 0". This will disable the X screen blank.
Gary Masters
From: Kragen Javier Sittler kragen@pobox.com
Check out the description of what doubleclick.net does at http://www.doubleclick.net/frames/adinfo/dartset.htm
Then decide whether you want to be added to their database of Internet user profiles. If not, you can use the script below; I run it in my
/etc/rc.d/rc.inet1. It prevents any DoubleClick banners from being displayed, prevents any cookies from being set, and prevents DoubleClick from collecting any data on you.
It also does the same thing with linkexchange.com, because I find their constant banners and requests for cookies annoying. If you'd prefer, you can take out the linkexchange lines.
However, this will also keep you from receiving *any* information from doubleclick or linkexchange directly... so you can't visit their web sites either.
On my machine, I put the script in
/etc/rc.d/rc.doubleclickand run it from
/etc/rc.d/rc.inet1at boot time, so I'm always protected from DoubleClick.
# Script begins below: #!/bin/sh # By Kragen Sitaker, 21 April 1997. # Prevent any packets from reaching doubleclick.net /sbin/route add -net 199.95.207.0 netmask 255.255.255.0 lo /sbin/route add -net 199.95.208.0 netmask 255.255.255.0 lo # And ad.linkexchange.com too! /sbin/route add -net 204.71.189.0 netmask 255.255.255.0 lo
From: Kidong Lee kidong@shinbiro.com
When I mount/umount file, I have to login as root. It's not convenient for me & other users. but, I found the solution that user who is not root can do mount/umount in mount man page.
Take a look at /etc/fstab.
#/dev/hdb /cdrom iso9660 ro,user 0 0
Note "user" in options field. In options field, if you add "user", users can do mount/umount.
From: Gregor Gerstmann, gerstman@tfh-berlin.de
Regarding Linux Gazette issue16, April 1997, I have some remarks
regarding the article on file transfer with the z protocol:
'I type sz
From: Walter Harms, Walter.Harms@Informatik.Uni-Oldenburg.DE Working on several different networks means that you always
need to copy your data from net to net. Most ppl use rcp but like most SysOps
I found this to be a terrible security hole. So as I started this job my first
business was to rewrite several scripts that were using rsh,rcp etc.
I replaced them with an ftp based script ftp - <input> out 2> out.err.
It's easy to
see that this was not a good idea because ftp was not intended as shell-commando
like cp,mv and the other guys. So I was happy to find the ftplib on a
linux-CD. It's a nice lib that I used to build cmds like ftpmv, ftpcp, ftprm..
This made my scripts much slimmer and simpler. I have some terrible
copy-scripts running but no problems copying on different systems
like Ultrix or AIX.
Example using ftpget (from the ftplib Author Thomas Pfau)
Who needs ftplib? Why use ftplib? Any drawbacks?
-- walter
Here is something interesting which you might consider for publication.
It is a short program written in LEX and C, which takes ASCII-Artwork
and translates it into HTML 3.0 compliant table data.
It is a pretty interesting idea, and as far as I know, I'm the first
person to try something like this, or automate the process.
The translator (a2t) has a few options:
The program was completed just today, so it is very new. I've
released it under the GNU license agreement.
For some examples of the output generated by a2t, see:
http://wilkes.edu/~pkeane
Enjoy--
Patrick
From: Martin Michlmayrtbm@cyrius.com
An example:
You can make references to the figure with
PostScript is already supported and the developer version of SGML-Tools
(the successor of Linuxdoc-SGML) now supports HTML as well. You
can specify a PostScript and a GIF file and depending on the output
(TeX or HTML) the respective image will be included.
Date: Wed Apr 2 12:15:54 1997 If you get sufficiently tweaked by the X monitor config problems,
I
suggest X Inside's AcceleratedX package. Its much simpler to configure
than the XFree package for both cards and monitors. I used to work
for them, but haven't in over a year. I still use their package because
its the easiest to handle all the video card/monitor details.
BTW, the monitor setup is menu based. If your monitor is not listed
you can just use one of the multisync if single frequency generic
configs. No dot clocks, but you do need to no your monitors frequency
capabilities. These should be listed in the monitors cdocumetntation.
The package is a commercial distribution and runs about $100 (last time
I checked). They change their name to Xi Graphics recently and the domain
for xinside.com might not be working right now. Try http://www.xig.com.
--
Michael J. Hammel
Date: Wed Apr 2 13:38:08 1997 Setting up the software is probably fairly straight forward. I've
never used MetroX (I use AcceleratedX instead), however. Basically
you'll have two choices:
The second choice is the one you need if you want to move the mouse
between the two monitors - like when the mouse goes past the right edge
of the first monitor it shows up on the left edge of the second monitor.
You'll have to check with Metro to find out which of these options is
supported and how to configure for it.
The hardware problem is tougher. The problem lies in the fact that PC's
were not originally designed with the idea that multiple display
adapters would be installed. The BIOS looks for an adapter at certain
locations (IRQ, I/O address) and, unless the second card is configurable
to some other address, the system will find multiple cards. What
happens next is in-determinant. Some systems won't boot. Some do but
don't display to either monitor correctly.
The trick is to find video adapters that were designed to be used
in conjunction with other video adapters. Many are not. The easiest
way for you to find out is check with Metro about what combinations of
video adapters they know work together. Chances are good the ones you
have don't. I know X Inside had a list of cards they knew work
together. You could search their web site (http://www.xinside.com or
http://www.xig.com) and see if that info is still there.
Hope this helps.
--
Michael J. Hammel
Date: Wed Apr 2 13:27:40 1997 Horse hockeys. 16 bit color is a limitation of the video subsystem and
has nothing to do with the memory of your system. Linear addressing in
the XFree86 X servers might be tied to system memory amounts, but that
would be a limitation in the XFree86 server, not in X. X defines
"method without policy", so such limitations just aren't built into X.
A couple of things you should note: The number of colors available
under 16bit displays is actually *less* than the number available to
8bit displays. Why this is true has to do with the way 16bit display
hardware works. The actual color palette for 8 bit displays can have
millions of colors - it can only display 256 colors at a time, however.
Frugal use of colormaps can allow you to have nearly exactly the right
colors for any given application. 16 bit displays only have a palette
of 65k (roughly) colors. Once those are used up, you're outta luck.
I'm not completely clear on what makes this difference such a problem
but if you visit the Gimp User's mailing list (see the Linux Graphics
mini-howto: http://www.csn.net/~mjhammel/linux/lgh.html) and ask this
question you'll get similar replies. Its been discussed quite at length
on the developers list, and most of them read the User's list.
BTW, if you want to see if Linear Addressing is the real problem, try
the X Inside AcceleratedX demo server and see if it works in 16 bit
color for you. Generally, your video card needs at least 1M of on board
RAM (not system memory - this is video memory on the video card) to
run in 16Bit mode, but then you'll probably only be able to run in
640x480 or (at most) 800x600 resolution. To run at higher resolutions
you'll need more video memory.
Hope this helps.
--
Michael J. Hammel
Date: Fri, 04 Apr 1997 13:20:40 -0600 On my Red Hat 4.0 system, the /etc/rc.d directory tree is where
everything happens. There are a lot of shell scripts in this set of
directories that are run when the system boots. To give yourself a
little more info, add some echo statements to the files. For example:
Now when the system is booting you can see exactly when rc.sysinit is
run, and what programs it launches. Repeat the above process for all the
scripts you find.
Now if the system hangs or gives an error during bootup you have a
better idea of where to look. If you don't have any problems while
booting then at least you have more info about what Linux is doing.
David
Date: Thu, 3 Apr 1997 20:38:02 +0300 (EET DST) Just replace the ppp.* with whatever you want (if you have multiple
ppps running). The easiest thing would to be write a script called ftphome
(or similar) and make it first assign the address and then doing ftp or
ncftp $ADDRESS. The snippet is originally from a local firewall, at the part
where it needs to know what its' own address is. :-)
A friend of mine at
mstr@ntc.nokia.com wrote this for me.
--
Kaj J. Niemi
Linux Expo, Research Triangle Park, NC (April 4, 1997) -
Enhanced Software Technologies, Inc. announced today that Groups of Linux
Users Everywhere (GLUE) will provide a free copy of the new BRU 2000 backup
and restore utility to GLUE user groups.
Enhanced Software Technologies, Inc. has joined Linux International as a
corporate member and is also offering members of GLUE user groups a
10-percent discount on purchases of BRU 2000.
Enhanced Software Technologies, Inc., a privately held corporation based in
Tempe, Arizona, is a leading provider of high-reliability systems
Additional information on BRU Giveaway.
GLUE is a project of SSC
publishers of Linux Journal. GLUE was implemented
to provide a world-wide member group for Linux User Groups. GLUE member
groups receive a subscription, materials for promoting and developing
their group, a way of advertising their group in a global setting, list-serv
and Linux Group location services, and discounts and samples from SSC and
Linux Journal. Other vendors may also offer special services or
discounts to GLUE member groups.
Additional information on Glue.
Solid Information Technology Ltd today announced a campaign
targeted at the community of Linux developers. Between March
and September 1997 Linux enthusiasts will be presented with
a free personal version of the robust SQL database engine
SOLID Server.
SOLID Server is a unique product by Solid Information Technology
Ltd, a privately held innovator of database
technologies.
To download your own copy of SOLID Desktop for Linux, access
http://www.solidtech.com/linuxfre.htm to find a site near you.
Additional Information:
The Elsop Webmaster Resource Center Contains links and comprehensive
coverage of computer industry trade publications,
website development, HTML, servers, validators,
link checkers, and software for webmasters.
Major sections include:
Produced and Sponsored by the
Electronic Software Publishing Corporation
http://www.elsop.com/linkscan
Do you consider yourself witty? Do you want to show your fabulous sense of
humor to the world? NOW IS YOUR CHANCE!
For several years now Linux Journal has been considering adding a monthly cartoon to
our magazine. We know who we could have "draw" the cartoons, but we really
don't have any idea what the jokes should be.
Please contribute any ideas you have for "Linux related" cartoons. The type
of cartoon we are imagining are one panel cartoons akin to what they have in
magazines like the New Yorker.
So send us your
favorite Linux jokes (one liners are best), and we will turn them into
cartoons.
For a good time, check out this website!
http://www.lightlink.com/fors/press/net-history.txt
There is a new user's group for Linux in Knoxvill, TN
They are called the Knoxville's Linux Users Base. Check out the web page
at
http://klub.ml.org
Take a look at the AfterStep Themes
page! Trae Mc Combs has been devoting some time to creating themes for
http://www.mindspring.com/~xwindow/as.html
Software Development Corporation http://www.sdcorp.com is working on
releasing version 7 of Corel's WordPerfect for Linux. It's expected to
ship sometime in April, with beta testing currently taking place.
Their webpages seem to warn that only beta testers have access to the
software, but following the links takes you to the download area where
they're freely available.
Here is a URL that has some interesting data:
http://fampm201.tu-graz.ac.at/karl/timings30.html 8 of the 10 fastest machines are running the Mac OS! the first windows
machine doesn't make a showing until 11th place( a pentium pro 200Mhz
running Windows NT 4.0) Incidently this ppro 200 is beat by a Mac 7500
150 Mhz!
You might wonder how this can be when the SPECint95 for Pentium Pros and
for Power PC 604's are so close? Its the operating system dummy!
What do I mean?
The Intel machines and the Macs are pretty equal, its Windows that slows
things down. If you check out the URL you'll see that although 8 of the
top ten are Macs or Mac clones, 2 of them are Intel pentium Pro 200Mhz
machines. Sadly for the Mac, the number one spot is a Pentium Pro 200
with 64 Meg RAM and a 256kb L2 cache running LINUX 2.0.27.
This barely beats the number 2 machine, a 225Mhz Power Tower Pro from
Power Computing with 256 Meg RAM and a 1Meg L2 cache.
The other Intel in the top 10 is a Pentium Pro 200Mhz with 128Meg Ram and
256Kb L2 cache, running NeXT STEP 3.3.
I don't think that Mac owners should be ashamed of losing to a LINUX
machine. LINUX is the result of an amazing effort put forth by many
dedicated programmers to produce a state of the art 32bit operating
system that utilizes hardware to the fullest. Mac users should be happy
that they can go head to head with such an OS, and still maintain the
great human interface of the Mac!
The only other contender is a NeXT machine! Wait'll your windows friends
see redbox!
Oh, BTW the first Win '95 machine doesn't make a showing until 15th
place. its a Pentium pro 200, 64 MB, 256kb, OS: Win95 and is just below a
PowerMac 7600/120, 48MB, 256kb, MacOS!
So if a windows user tells you their machine is faster, tell them that
you know...if they switch to LINUX.
The development of 'wp', a word processor for the Linux environment has
recently been started. Although it's primary goal is a Linux-based word
processor, wp will eventually be available for many other platforms.
WP is an open system, object orientated, and object driven; written mainly
using C++, although little code has of yet been written. The current
objective is a full design specification/mission statement and determining
the current products that can be used to help the development of the
product further.
Because of this openness, it is proposed to have the user interface
seperate from the main program; the reason for this meaning that the user
can choose whichever interface suits them best, from a ncurses driven text
interface to an X-Windows display using different widget sets.
The web site for Wp is at http://sunsite.unc.edu/paulc/wp
If you wish to obtain the design specification notes for wp, they are also
available at the above site.
A FAQ is currently being prepared, if you have any questions or
suggestions, please send them to wp@squiznet.demon.co.uk
If you wish to contribute to the project in any form, please contact
paulc@sunsite.unc.edu and introduce yourself, a copy of which will be sent
to the wp-developers mailing list unless you specifically state that you
do not wish for this to happen.
Xcoral-3.0 has been released and now available on the Net.
Xcoral is a multiwindow mouse-based text editor for
the X Window System. It contains a built-in browser that enables you to
navigate through C functions, C++ classes, Java classes, methods and files.
It also contains a SMall Ansi C Interpreter (Smac) which is also built-in
to extend the editor's possibilities
(user functions, key bindings, modes etc).
Xcoral provides variable width fonts, menus, toolbar,
scrollbars, buttons, search, regions, kill-buffers, macros and
undo. An on-line manual box, with a table of contents
and an index, helps you to use and customize the editor.
Xcoral also offers facilities to write Latex documents
and Html pages.
Xcoral is a direct Xlib client and runs on color/bw
X Display.
OS: SunOS 4.1.x, Solaris 2.[45], LINUX, AIX, HPUX,
IRIX and OSF-1.
Changes from xcoral-2.5:
The Linux/Alpha team at Digital Equipment Corporation
today is releasing a developers' beta version of EM86, a Linux/x86
emulator for Linux/Alpha. Using components of the
DIGITAL FX!32 technology, EM86 is a software emulator that enables
Linux/Alpha systems to run Linux/x86 software without modification.
EM86 currently supports statically linked and dynamically linked x86
ELF32 binaries under Linux/Alpha. Future enhancements will include
support for iBCS-2 compliant executables, improved emulator performance,
and interoperation with native Alpha code. A release incorporating
these features is anticipated in July, 1997.
They are releasing a beta version of EM86 at this time to provide
Linux developers early access to the software, to aid in the
verification of software packages, and to provide feedback and bug
reports to the Linux/Alpha team.
The following Linux/x86 software packages run successfully on this
beta version of EM86, with some qualifications as described in the
README file included in the distribution:
EM86 may be obtained via anonymous ftp from:
ftp://ftp.digital.com/pub/DEC/Linux-Alpha/em86
XForms V0.86 is now available from:
XForms is a graphical user interface toolkit and builder based on Xlib
for X Window Systems. XForms is a portable and efficient C library
that can be used in both C and C++ programs.
The library works in all visuals and all depths (1-24) and comes
with a rich set of objects such as buttons (of many flavors, including color
XPMs as labels) , browsers, sliders, and menus integrated into an
elegant event/object callback execution model that allows fast and
easy construction of X-applications. It also has OpenGL (on SGI) and
Mesa support.
XForms comes bundled with
perl, ada95, python and fortran bindings to xforms are in alpha/beta.
Please visit the xforms' home page for more info.
Debian 1.3 is now in beta test. We are performing a month-long test
with an organized quality control team. If you'd like to be an official
beta tester, please contact Dale Scheetz dwarf@polaris.net .
The Debian 1.3 files are under the "frozen" directory on most of the
Debian mirror sites. There are now 73 Debian mirrors worldwide! You can
find the mirror list at
ftp://ftp.debian.org/debian/README.mirrors or
ftp://debian.crosslink.net/pub/debian/README.mirrors. Please consider
that this is beta-quality software and there will be bugs. If you have
any problem, please see the information on our bug-tracking system
at http://www.debian.org/support.html, or write to Dale at the above
address.
Announcing the public availability of the Freedom
Desktop Lite. Freedom Desktop Lite is a
desktop environment/GUI integrated to the Unix
environment. It
helps users interact with Unix quickly and efficiently. Freedom
Desktop runs transparently in a variety of Unix environments,
from Desktop computers (i.e. Linux) to enterprise workstations.
The Freedom Desktop Lite environment bundles the following
applications:
From: Aaron M. Lee aaron@shifty.adosea.com Howdy Jim,
My name's Aaron and I am sysadmin Cybercom Corp., an ISP in
College Station, TX. We run nothing but Linux, and have been involved w/
a lot of hacking and development on a number of projects. I have an
unusual problem and have exhausted my resources for finding an answer- so
I thought you might be able to help me out, if you've got the time.
Anyway, here goes...
I've got a scsi disk I was running under Sparclinux that has 3
partitions, 1 Sun wholedisk label, 2 ext2. That machine had a heart
attack, and we don't have any spare Hypersparcs around- but I _really_
need to be able to mount that drive to get some stuff off of it. I compiled
in UFS fs support w/ Sun disklabel support into the kernel of an i386 Linux
box, but the when I try to mount it, it complains that /dev/sd** isn't a
valid block device, w/ either the '-t ufs' or '-t ext2' options. Also,
fdisk thinks the fs is toast, and complains that the blocks don't end
in physical boundaries (which is probably the case for an fdisk that
doesn't know about Sun disklabels), and can't even tell that the
partitions are ext2 (it thinks one of them is AIX!). Any ideas?
Consider the nascent state of Sparc support for Linux
I'm not terribly surprised that you're having problems.
You seem to be asking:
"How do I get Linux/Intel to see the fs on
this disk?"
However I'm going to step back from the that question
and ask the broader question:
"How do you recover the (important) data off of
that disk in a usable form?"
Then I'll step back even further and ask:
"How important is that data? (what is its
recovery worth to you)?"
... and If you are like most ISP's out there -- you have not
disaster or recovery plans, and little or no backup
strategy. Your boss essentially asks you to running
back and forth on the high wire at top speed -- without
a net.
As a professional sysadmin you must resist the pressure
to perform in this manner -- or at least you owe it to
yourself to carefully spell out the risks.
In this case you had a piece of equipment that was
unique the Sparc system -- so that any failure of
any of its components would result in the lack of
access to all data on that system.
Your question makes it clear that you didn't have
sufficiently recent backups of the data on that
system (otherwise the obvious solution would be
to restore the data to some other system and
reformat the drive in question).
My advice would be to rent (or even borrow) a
SPARC system for a couple of days (a week is a
common minimum rental period) -- and install
the disk into that.
Before going to the expense of renting a system
(or buying a used one) you might want to ensure
that the drive is readable at the lowest physical
level. Try the dd command on that device. Something
like:
... should let you know if the hardware is operational.
If that doesn't work -- double and triple-check all of the
cabling, SCSI ID settings, termination and other hardware
compatibility issues. (You may be having some weird problem
with a SCSI II differential drive connecting to an
incompatible controller -- if this is an Adaptec 1542B
-- be sure to break it in half before throwing it away
to save someone else the temptation (the 1542C series is
fine but the B series is *BAD*)).
Once you are reasonably confident that the hardware
is talking to your system I'd suggest doing a direct,
bitwise, dump of the disk to a tape drive. Just use a
command like:
... if you don't have a sufficiently large tape drive
(or at least a sufficiently large spare hard disk) *and
can't get one* than consider looking for a better
employer.
Once you have a tape backup you can always get back
to where you are now. This might not seem so great
(since you're clearly not where you'd like to be) but
it might be infinitely preferable to where you'll be
if you have a catastrophic failure on mounting/fsck'ing
that disk.
For the broader problem (the organizational ones rather
than the technical ones) -- you need to review
the requirements and expectations of your employer --
and match those against the resources that are being
provided.
If they require/expect reliable access to their data --
they must provide resources towards that end. The most
often overlooked resource (in this case) is sysadmin
time and training. You need the time to develop
disaster/recovery plans -- and the resources to test
them. (You'd be truly horrified at the number of sites
that religiously "do backups" but have an entire staff that
has never restored a single file from those).
Many organizations can't (or won't) afford a full spare
system -- particularly of their expensive Sparc stations.
They consider any system that's sitting on a shelf to be a
"waste." -- This is a perfectly valid point of view.
However -- if the production servers and systems are
contributing anything to the companies bottom line --
there should be a calculable cost for down time. If that's
the case then there is a basis for comparison to the costs of
rentals, and the costs of "spare" systems.
Organizations that have been informed of this risks and
costs (by there IS staff) and continue to be unwilling or
unable to provide the necessary resources will probably
fail.
Thanks in advance for any possible help,
--Aaron
It's often the case that I respond with things that
I suspect my customer don't want to hear.
The loss of this data (or the time lost to recovering
it) is an opportunity to learn and plan -- you may
prevent the loss of much more important information
down the road if you now start planning for the
inevitable hardware and system failures.
From:Steven W., steven@gator.net Can you help me? Do you know of a Unix (preferably Linux) emulator
that runs under Windows95?
-- Steven.
Short Answer:
I don't know of one.
Longer Answer:
This is a tough question because it really doesn't
*mean* anything. An emulator is a piece of software
that provide equivalent functionality to other software
or hardware. Hopefully this software is indistinguishable
from the "real" thing in all ways that count.
(Usually this isn't the case -- most VT100
terminal emulation packages have bugs in them
-- and that is one of the least complicated
and most widespread cases of emulation in the
world).
A Unix "emulator" that ran under Win '95 would probably not be
of much use. However I have to ask what set of features
you want emulated?
Do you want a Unix-like command shell (like
Korn or Bash)? This would give you some of the
"feel" of Unix.
Do you want a program that emulates one of the
GUI's that's common on Unix? There are X Windows
"display servers" (sort of like "emulators") that
run under NT and '95. Quarterdeck's eXpertise
would be the first I would try.
Do you want a program that allows you to run
some Unix programs under Win '95? There are
DOS, OS/2, and Windows (16 and 32 bit) ports of
many popular Unix programs -- including most of
the GNU utilities. Thus bash, perl, awk, sed,
vi, emacs, tar, and hundreds of other utilities
can be had -- most of them for free.
Do you want to run pre-compiled Unix binaries
under Win '95? This would be a very odd request
since there are dozens of implementations of
Unix for the PC platform and hundreds for other
architectures (ranging from Unicos on Cray super-
computers to Minix and Coherent on XT's and 286's).
Binary compatibility has playing only a tiny role
in the overall Unix picture. I suspect that
supporting iBCS (a standard for Unix binaries on
intel processors -- PC's) under Win '95 would be a
major technical challenge (and probably never
provide truly satisfying results).
*note*: One of the papers presented at Usenix in
Anaheim a couple of months ago discussed the
feasibility of implementing an improved Unix
subsystem under NT -- whose claim of POSIX support
as proven to be almost completely useless in the
real world. Please feel free to get a copy of
the Usenix proceeding if you want the gory details
on that. It might be construed as a "Unix emulation"
for Windows NT -- and it might even be applicable to
Win '95 -- with enough work.
If you're willing to run your Windows programs
under Unix there's hope. WABI currently supports
a variety of 16-bit Windows programs under Linux
(and a different version support them under Solaris).
Also work is continuing on the WINE project -- and
some people have reported some success in running
Windows 3.1 in "standard mode" under dosemu (the
Linux PC BIOS emulator). The next version of WABI
is expect to support (at least some) 32-bit Windows
programs.
My suggestion -- if this is of any real importance to you --
is that you either boot between Unix and DOS/Windows or that
you configure a separate machine as a Unix host -- put it in
a corner -- and using your Win '95 system as a terminal,
telnet/k95 client and/or an X Windows "terminal" (display
server).
By running any combination of these programs on your Windows
box and connecting to your Linux/Unix system you won't have
to settle for "emulation." You'll have the real thing --
from both sides. In fact one Linux system can serve as the
"Unix emulation adapter" for about as many DOS and Windows
systems as you care to connect to it.
(I have one system at a client site that has about 32Mb
of RAM and 3Gb -- it's shared by about 300 shell and
POP mail users. Granted only about 20 or 30 of them are
ever shelled at any given time but it's no where near it's
capacity).
I hope this gives you some idea why your question is
a little non-sensical. Operating systems can be viewed
from three sides -- user interface (UI), applications
programming interface (API), and supported hardware
(architecture).
Emulating one OS under another might refer to emulating
the UI, or the API or both. Usually emulation of the
hardware support is not feasible (i.e. we can't run DOS
device drivers to provide Linux hardware support).
If one implemented the full set of Unix system calls
in a Win '95 program that provided a set of "drivers"
to translate a set of Unix like hardware abstractions
into calls to the Windows device drivers -- and one
ported a reasonable selection of software to run under
this "WinUnix kernel" -- one could call that "Unix emulation."
However it would be more accurate to say that you had
implemented a new version of Unix on a virtual machine
which you hosted under Windows.
Oddly enough this is quite similar to what the Lucent
(Formerly Bell Labs?) Inferno package does. Inferno
seems to have evolved out of the Plan 9 research project
-- which apparently was Dennis Ritchie's pet project for
a number of years. I really don't know enough about
the background of this package -- but I have a CD
(distributed to attendees of the aforementioned Usenix
conference) which has demo copies of Inferno for several
"virtual machine" platforms (including Windows and Linux).
Inferno is also available as a "native" OS for a couple
of platforms (where it includes it's own device drivers
and is compiled as direct machine code for a machine's
platform).
One reason I mention Inferno is that I've heard that
it offers features and semantics that are very similar
to those that are common in Unix. I've heard it described
as a logical outgrowth of Unix that eschews some of the
accumulation of idiosyncrasies that has plagued Unix.
One of these days I'll have to learn more about that.
I have Windows95 and Linux on my system, on separate partitions, I
can't afford special equipment for having them on separate machines.
I really like Linux, and Xwindows, mostly because of their great
security features. (I could let anybody use my computer without
worrying about them getting into my personal files). Windows95's
pseudo-multi-user system sucks really bad. So, mainly, this is why I
like Linux. I also like the way it looks. Anyways, I would just run
Linux but my problem is that Xwindows doesn't have advanced support
for my video card, so the best I can get is 640x480x16colors and I
just can't deal with that. Maybe I'm spoiled. The guy I wrote on
the Xwin development team told me that they were working on better
support for my card, though. (Aliance Pro-Motion). But, meanwhile,
I can't deal with that LOW resolution. The big top-it-off problem is
that I don't know of anyway to have Linux running _while_ Win95 is
running, if there even is a way. If there was, it would be great,
but as it is I have to constantly reboot and I don't' like it. So
this is how I came to the point of asking for an emulator. Maybe
that's not what I need after all. So what can I do? Or does the
means for what I want not exist yet?
-- Steven.
If you prefer the existing Linux/X applications and
user interface -- and the crux of the problem is support
for your video hardware -- focus on that. It's a simpler
problem -- and probably offers a simpler solution.
There are basically three ways to deal with a lack of
XFree86 support for your video card:
Be sure to contact the manufacturer to ask for a
driver. Point out that they may be able to make
small changes to an existing XFree86 driver. You
can even offer to help them find a volunteer
(where you post to the comp.os.linux.dev...sys.
newsgroup and one or two of the developer's mailing
lists -- and offer some support). Just offering to
do some of the "legwork" maybe be a significant
contribution.
This is an opportunity to be a "Linux-Activist."
--
Jim
From:Charles A. Barrassocharles@blitz.com
I'm sure I gave a lengthy answer to this fairly recently.
Maybe it will appear in this month's issue (or maybe
I answered it on a newsgroup somewhere).
In any event, the short answer is: You don't.
The PC architecture doesn't support using multiple
VGA/EGA cards concurrently. I don't think XFree86 can
work with CGA cards (and who'd want to!). You might
be able to get a Hercules compatible Monochrome Graphics
Adapter (MGA) to work concurrently with a VGA card (since
they don't use overlapping address spaces). I don't know
if this is the method that Metro-X supports.
There are specialized video adapters (typically very expensive
-- formerly in the $3000+ range) that can co-exist with
VGA cards. Two sets of initials that I vaguely recall are
TIGA and DGIS. Considering that you seem unwilling to
pay $100 (tops) for a copy of Metro-X I think these --
even if you can still find any of them -- are way out of
your price league.
Another, reasonable, alternative is to connect a whole
Xterminal or another whole system and run X on that. You
can then remotely display your windows on that about as
easily as you could set them to display on the local
server.
(I know -- you might not get some cool window manager
to let you drag windows from one display server to another
-- a trick which I've seen done with Macs under MacOS and
with Suns and SGI's. But I've never set one of those up
anyway -- so I couldn't begin to help you there).
You might double check with the Metro-X people to see
what specific hardware is required/supported by their
multiple display feature and then check with the XFree86.org
to see if anyone has any drivers for one of those supported
configurations.
As a snide note I find your phrase "that costs money :("
to be mildly offensive. First the cost of an additional
monitor has got to be at least 3 times the price of
a copy of Metro-X. Second "free" software is not about
"not having to pay money."
I'm not trying to sell you a copy of Metro-X here. I
don't use it -- and I specifically choose videos cards
that are supported by XFree86 when I buy my equipments.
Likewise I don't recommend Linux to my customers because
it "doesn't cost them anything." In fact it does cost
them the time it takes me to install, configure and maintain
it -- which goes for about $95/hr currently. I recommend
Linux because it is a better tool for many jobs -- and because
the benefits of it's being "free" -- in the GNU sense of the
term -- are an assurance that no one can "have them over a
barrel" for upgrades or additional "licensing" fees. They are
always *free* to deploy Linux on as many systems as they want,
have as many users and/or processes as they want on any system,
make their own modifications to the vast majority of tools
on the system or hire any consultants they want to make the
customizations they need.
I'm sorry to be so "political" here -- but complaining
that Metro-X "costs money" and asking me for a way to
get around that just cost me about $50 worth of my time.
Heck -- I'll go double or nothing -- send my your postal
address and I'll buy you a copy of RedHat 4.1. That comes
with a license for one installation of Metro-X and only
costs about $50. I'll even cover the shipping and handling.
(Please call them first to make sure that it really does
support your intended hardware configuration).
Thanks for the time,
No problem. (I did say "mildly" didn't I).
--
Jim
From: Wietse Venema wietse@szv.sin.tue.nl Wietse
Thanks for the quick response. I'll have to play with
that. I suppose a custom "virtual finderd" would
be a good experiment.
Do you know where there are any working examples of this
and the twist option posted to the 'net? I fight with
some of these and don't seem to get the right results.
What I'd like is an example that drops someone into a
chroot'd jail as "nobody" or "guest" and running a
copy of lynx if they are from one address -- but
lets them log in a a normal user if they are from an
internal address. (We'll assume a good anti-spoofing
packet-filter on the router(s)).
Did you ever add the chrootuid functionality to tcpd?
How would you feel about an option to combine the
hosts.allow and hosts.deny into just tcpd.conf?
(I know I can already put all the ALLOW and DENY
directives in a single file -- and I'm not much of a
programmer but even *I* could patch my own copy to
change the filename -- I'm just talking about the
general case).
SERVER ENDPOINT PATTERNS (which is what he said one to me when I suggested merging
his chrootuid code with tcpd).
I've blind copied Wietse on this (Hi!). I doubt he has
time to read the Linux Gazette.
--
Jim
From:Wietse Venema, wietse@wzv.win.tue.nl
Use "twist" to run a service that depends on destination address:
fingerd@host1: ALL: twist /some/where/fingerd-for-host1
What I'd like is an example that drops someone into a
chroot'd jail as "nobody" or "guest" and running a
copy of lynx if they are from one address -- but
lets them log in a a normal user if they are from an
internal address. (We'll assume a good anti-spoofing
packet-filter on the router(s)).
I have a little program called chrootuid that you could use.
Did you ever add the chrootuid functionality to tcpd?
I would do that if there was a performance problem. Two small
programs really is more secure than a bigger one.
How would you feel about an option to combine the
hosts.allow and hosts.deny into just tcpd.conf?
What about compatibility with 1 million installations world-wide?
(I know I can already put all the ALLOW and DENY
directives in a single file -- and I'm not much of a
programmer but even *I* could patch my own copy to
change the filename -- I'm just talking about the
general case).
This is because the language evolved over time. Compatibility can
become a pain in the rear.
--
Weitse
From:Kenneth Ng, kenng@kpmg.com And that's it. Granted ssh is better. But sometimes you have to go
somewhere that
only supports ftp.
That's one of several ways. Another is to use ncftp
-- which supports things like a "redial" option to keep
trying a busy server until it gets through. ncftp also has
a more advanced macro facility than the standard .netrc (FTP).
You can also use various Perl and Python libraries (or classes)
to open ftp sessions and control them. You could use 'expect'
to spawn and control the ftp program.
All of these methods are more flexible and much more robust
than using the standard ftp client with redirection ("here"
document or otherwise).
--
Jim
From: Stephen P. Smith, ischis@evergreen.com If I try an anonymous ftp session, the email password is rejected.
what are the possible sources of failure?
where should i be going for more help? :-)
Do you have a user named 'ftp' in the /etc/passwd file?
done.
wu-ftpd takes that as a hint to allow *anonymous* FTP.
If you do have one -- or need to create one -- be sure that
the password for it is "starred out." wu-ftpd will not
authenticate against the system password that's defined for a
a user named "ftp."
done.
You should also set the shell to something like /bin/false or
/bin/sync (make sure that /bin/false is really a binary and
*not* a shell script -- there are security problems -- involve
IFS (inter-field separators) if you use a shell script in the
/etc/passwd shell field).
done.
There is an FAQ for anonymous FTP (that's not Linux specific).
There is also a How-To for FTP -- that is more Linux oriented.
If you search Yahoo! on "wu-ftp" you'll find the web pages
at Washington University (where it was created) and at
academ.com -- a consulting service that's taken over development
of the current beta's.
Guess I will just have to do it the hard
way. Will tell you what I find (just in
case you want to know.
What does your /etc/ftpaccess file look like?
Did you compile a different path for the ftpaccess file
(like /usr/local/etc/)?
What authentication libraries are you using (old
fashioned DES hashes in the /etc/passwd, shadow,
shadow with MD5 hashes -- like FreeBSD's default,
or the new PAM stuff)?
Is this invoked through inetd.conf with tcpd
(the TCP Wrappers)? If so, what does your /var/log/messages
say after a login failure? (Hint: use the command:
'tail -f /var/log/messages > /dev/tty7 &' to leave a continuously
updated copy of the messages file sitting on one of your
-- normally unused -- virtual consoles).
One trick I've used to debug inetd launched programs (like
ftpd and telnetd) is to wedge a copy of strace into the
loop. Change the reference to wu.ftpd to trace.ftpd --
create a shell or perl script named trace.ftpd that consists
of something like:
... and then inspect the strace file for clues about
what failed. (This is handy for finding out that the
program couldn't find a particular library or configuration
file -- or some weird permissions problems, etc).
--
Jim
From: Yash Khemani, khemani@plexstar.com i am guessing that the lilo floppy does not have on it the pcmcia
drivers. what is the solution at this point to run RedHat on this
machine?
You've got the right idea.
The 1010101010101... from LILO is a dead giveaway that
your kernel is located on some device that cannot be
accessed via the BIOS.
There are a couple of ways to solve the problem.
I'd suggest LOADLIN.EXE.
LOADLIN.EXE is a DOS program (which you might have
guessed by the name) -- which can load a Linux kernel
(stored as a DOS file) and pass it parameters (like
LILO does). Basically LOADLIN loads a kernel (Linux or
FreeBSD -- possibly others) which then "kicks" DOS
"out from under it." In other words -- it's a one-way
trip. The only way back to DOS is to reboot (or
run dosemu ;-) .
LOADLIN is VCPI compatible -- meaning that it can run
from a DOS command prompt even when you have a memory
manager (like QEMM) loaded. You can also set LOADLIN
as your "shell" in the CONFIG.SYS. That's particularly
handy if you're using any of the later versions of DOS
that support a multi-boot CONFIG.SYS (or you're using the
MBOOT.SYS driver that provided multi-boot features in
older versions of DOS).
To use LOADLIN you may have to create a REALBIOS.INT
file (a map of the interrupt vectors that are set by
your hardware -- before any drivers are loaded).
To do this you use a program (REALBIOS.EXE) to create
a special boot floppy, then you boot off that floppy
(which records the interrupt vector table in a file)
-- reboot back off your DOS system and run the second
stage of the REALBIOS.EXE.
This little song and dance may be necessary for each
hardware configuration. (However you can save and
copy each of the REALBIOS.INT files if you have a
couple of configurations that you switch between --
say, with a docking station and without).
With LOADLIN you could create a DOS bootable floppy,
with a copy of LOADLIN.EXE and a kernel (and the
REALBIOS.INT -- if it exists). All of that will
just barely fit on a 1.44M floppy.
Another way to do this would be to create a
normal DOS directory on your laptop's IDE drive --
let's call it C:\LINUX (just to be creative).
Then you'd put your LOADLIN.EXE and as many different
kernels as you liked in that directory -- and maybe
a batch file (maybe it could be called LINUX.BAT) to
call LOADLIN with your preferred parameters. Here's a
typical LINUX.BAT:
(where LNX2029.KRN might be a copy of the Linux-2.0.29
kernel -- with a suitable DOS name).
I'd also recommend another batch file (SINGLE.BAT) that
loads Linux in single-user mode (for fixing things when
they are broken). That would replace the LOADLIN line
in the LINUX.BAT with a line like:
Another way to do all of this is to simply dd a
properly configured kernel to a floppy. You use the
rdev command to patch the root device flags in the
kernel and dump it to a floppy. This works because
a Linux kernel is designed to work as a boot image.
The only problem with this approach is that it doesn't
allow you to pass any parameters to your kernel (to
force single user mode, to select an alternate root
device/filesystem, or whatever).
For other people who have a DOS system and want to
try Linux -- but don't want to "commit" to it with
a "whole" hard drive -- I recommend DOSLINUX.
A while back there was a small distribution called
MiniLinux (and another called XDenu) which could
install entirely within a normal DOS partition --
using the UMSDOS filesystem. Unfortunately MiniLinux
has not been maintained -- so it's stuck with a 1.2
kernel and libraries.
There were several iterations of a distribution called
DILINUX (DI= "Drop In") -- which appears to have eventually
evolved into DOSLINUX. The most recent DOSLINUX seems was
uploaded to the Incoming at Sunsite within the last two
weeks -- it includes a 2.0.29 kernel.
The point MiniLinux and DOSLINUX is to allow one to install
a copy of Linux on a DOS system as though it were a DOS
program. DOSLINUX comes as about 10Mb of compressed
files -- and installs in about 20-30Mb of DOS file space.
It includes Lynx, Minicom, and a suite of other utilities
and applications.
All in all this is a quick and painless way to try Linux.
So, if you have a DOS using friend who's sitting on the fence,
give them a copy of DOSLINUX and show them how easy it is.
thanks!
You're welcome.
(Oh -- you might want to get those shift keys fixed --
e.e. cummings might sue for "look and feel")
--
Jim
I'm familiar with C-Kermit. In fact I may have an
article in the June issue of SysAdmin magazine on that very
topic.
The main points of my article are that C-Kermit is a
telnet and rlogin client as well as a serial communications
program -- and that it is a scripting language that's
available on just about every platform around.
I know about Telix' support for the kermit transfer protocol.
It sucks. On my main system I get about 1900 cps for
ZMODEM transfers -- about 2200 for kermit FAST (between
a copy of C-Kermit 5A(188) and 6.0.192 and about 70 cps
(yes -- seventy!) between a copy of C-Kermit and Telix'
internal kermit.
Other than that I've always liked Telix. Minicom has
nice ncurses and color -- but is not nearly as featureful
or stable as either Telix for DOS or any version of C-Kermit.
Your line hangups probably have to do with your settings for
carrier-watch. Try SET CARRIER-WATCH OFF or ON and see if
it still "hangs" your line. I suspect that its actually just
doing read() or write() calls in "blocking" mode. You might
have to SET FLOW-CONTROL NONE, too. There are lots of
C-Kermit settings. If you continue to have trouble -- post
a message to the comp.protocols.kermit.misc newsgroup
(preferred) or send a message to kermit-support@columbia.edu.
When I first started using C-Kermit (all of about two months
ago) my initial questions where answered by Frank da Cruz
himself (he's the creator of the Kermit protocol and the
technical lead of the Kermit project at Columbia University).
(That was before he knew that I'm a "journalist" -- O.K.
quit laughing!). Frank is also quite active in the newsgroup.
I think he provides about 70 or 80 per cent of the technical
support for the project.
Oh yeah! If you're using C-Kermit you should get the
_Using_C-Kermit_ book. It was written by Frank da Cruz and
Christine Gianone -- and is the principal source of funding
for the Kermit project. From what I gather a copy of the
book is your license to use the software.
--
Jim
From: Robert Rambo, robert.rambo@yale.edu
If using the 16 bit plan (16bpp) mode to increase
your color depth -- that suggests that selecting this
mode is causing the server to use a lower resolution.
That is completely reasonable. If you have a 2Mb video
card and you run it in 1024x768x256 or 1024x768x16 --
then you try to run it with twice as many colors --
the video RAM has to come from somewhere. So it
bumps you down to 800x600 or 640x480. These are just
examples. I don't deal with graphics much so I'd have
to play with a calculator to figure the actual maximum
modes that various amounts of video RAM could support.
There are alot of settings in the XConfig file. You
may be able to tweak them to do much more with your
existing video card. As I've said before -- XConfig
files are still magic to me. They shifted from blackest
night to a sort of charcoal gray -- but I can't do them
justice in a little article hear. Pretty much I'd have
to lay hands on it -- and mess with it for a couple of
hours (and I'm definitely not the best one for that job).
If you haven't upgraded to a newer XFree86 (3.2?) then
this would be a good time to try that. The newer one
is much easier to configure and supports a better selection
of hardware -- to a better degree than the older versions.
I haven't heard of any serious bugs or problems with
the upgrades.
You may also want to consider one of the commercial servers.
Definitely check with them in advance to be absolutely certain
that your hardware is supported before you buy. Ask around in
the newsgroups for opinions about your combination of hardware.
It may be that the XFree86 supports you particular card better
than Metro-X or whatever.
You may also want to look at beefing up your video hardware.
As I've said -- I don't know the exact figures -- but I'd
say that you probably need a 4Mb card for anything like
16bpp at 1024x768. You should be able to look up the
supported modes in your card's documentation or on the
manufacturer's web site or BBS.
Also, is there some way to change the color depth
setting to start X with a depth of 16 every time. I do not use the XDM
manager to initiate an X session.
Yes -- it's somewhere in that XConfig file. I don't
remember the exact line. I really wish a bona fide GUI
X wiz would sign up for some of this "Answer Guy" service.
It doesn't matter whether you use xdm or not. If you
put the desired mode in the XConfig file. However --
since you don't you could just write your own wrapper
script, alias or shell function to call 'startx' with
the -- -bpp16 options. You could even re-write 'startx'
(it is just a shell script). That may seem like cheating --
but it may be easier than fighting your way through the
XConfig file (do you get the impression that I just don't
like that thing -- it is better than a WIN.INI or a
SYSTEM.INI -- but not be much).
--
Jim Dennis,
From: Brian Moore, bem@thorin.cmc.net
Will these support the real features (storing and
organizing folders on the server side)?
I heard that NS "Communicator" (the next release
Netscape's Navigator series is apparently going to
come with a name change) supports IMAP -- but it's
possible to implement this support as just a variant
of POP -- get all the message and immediately
expunge all of them from the server.
It seems that this is how Eric S. Raymond's 'fetchmail'
treating IMAP mail boxes -- as of about 2.5 (it seems
that he's up to 3.x now)
The easiest IMAP server to install is certainly the University of
Washington server. It works, handles nearly every mailbox format around
and is very stable. It's also written by the guy in charge of the IMAP
spec itself, Mark Crispin.
As for clients, there is always Pine, which knows how to do IMAP quite
well. This is part of most Linux distributions as well.
I did mention pine. However it's not my personal favorite.
Do you know of a way to integrate IMAP with emacs mh-e/Gnus
(or any mh compatible folder management system)?
For GUI clients there is ML, which is a nice client, but requires Motif
and can be slow as sin over a modem when you have a large mailbox.
That's available in source at
http://www-CAMIS.Stanford.EDU/projects/imap/ml
I thought I mentioned that one as well -- but it's
a blur to me.
I personally avoid GUI's like the plague. I'm
typing this from my laptop, through a null modem link
to my machine in the other room.
I run emacs under screen -- so I can use mh-e for most
mail, Gnus for netnews and for some of my mailing lists
(it can show news folders as though they were threaded
news groups). screen allows me to detach my session from
my terminal so I can log out, take off with the laptop,
and re-attach to the same session later (via modem or when
I get back home).
Asking on the mailing list about static linked linux versions will get
you one (and enough nagging may get them to actually put one of the
current version up).
ML is really the nicest mail client I have ever used.
As for pop daemons with UIDL support, go for qpopper from qualcomm.
ftp.qualcomm.com somewhere. Has UIDL and works fine.
O.K. I'll at that to my list.
Does that one also support APOP's authentication
mechanism (which I gather prevents disclosing your
password over an untrusted network by using something
like an MD5 hash of your password concatenated with
a date and time string -- or something like that)?
Does qpopper allow you to maintain a POP user account
file that's separate from your /etc/passwd file?
Do you know of an IMAP server that supports these
sorts of features (secure authentication and separate
user base)?
(I know this probably seems like a switch -- the
so called "Answer Guy" asking all the questions --
but hey -- I've got to get my answers from *somewhere*)
--
Jim
From: Graham Todd, gtodd@yorku.ca
The Netscape's Communicator 4.0b2 thing does too but there are so many
other ugly bits that I'm not gonna bite.
Jeez pretty soon with this fancy new IMAP stuff you'll be able to do
almost as much as you can right now with emacs and ange-ftp (which I
use regularly to access remote mail folders and boxes with out having
to login - it's all set up in .netrc).
Of course the answer is almost always "emacs" .... BTW Linux
makes a GREAT program loader for emacs ;-)
Seems kind of kludgey. Besides -- does that
give you the main feature that's driving the creation
of the IMAP/ACAP standards? Does it let you
store your mail on a server and replicate that to
a couple of different machines (say your desktop and
your laptop) so you can read and respond to mail "offline"
and from *either* system?
Yeah, more or less. If you save the mail on your server to local
folders or make a local folder be /me@other.mail.host:/usr/spool/me.
Using ange-ftp to me seem exactly like IMAP in Pine or Netscape
communicator 4.0b2. Though apparently IMAP will update folders across
hosts so that only that mail deleted locally (while offline) will get
deleted on the remote host on the next login etc. etc. I don't know
much about IMAP's technical standard either but find I get equal mail
management capability from ange-ftp/VM. (equal to Pine and
Communicator so far).
WARNING: In a week or so when I get time I'm gonna ask you a tricky
question about emacs and xemacs.
Feel free. Of course I do know a bit more about emacs
than I do about X -- so you may not like my answer much.
Heh heh OK...
(comp.emacs.xemacs is silent on this). Emacs running as emacs -nw in
a tty (i.e console or an xterm) runs fine and lets me use all the job
control commands (suspend/fg etc) but with Xemacs job control won't
work unless I'm running as root. That is if I'm running "xemacs" or
"xemacs -nw" in an xterm or at the console and do C-z and then once
I'm done in the shell I do "fg", xemacs comes back but the keyboard
seems to be bound to the tty/console settings (Ctrl-z Ctrl-s Ctrl-q
etc all respond as if I were in a dumb terminal). The only recourse
is to Ctrl-z back out and kill xemacs. This does not happen if I run
xemacs setuid root (impractical/scary) or as root (scary). Something
somewhere that requires root permission or suid to reset the tty
characteristics doesn't have it in xemacs - but does in emacs...
My only response so far has been that "you'll have to
rebuild/recompile your xemacs" - but surely this wrong. Does anything
more obvious occur to you? I feel it must be something simple in my
set up (RH Linux 2.0.29). Of course if I could get this fixed I'd
start feeling more comfortable not having GNU-Emacs on my machine ;-)
which may not be an outcome you would favour.
I once had a problem similar to this one -- suspending
minicom would suspend the task and lock me out of it.
It seemed that the ownership of the tty was being
changed.
So -- the question comes up -- what permissions are set on
your /dev/tty* nodes. It seems that most Linux distributions
are set up to have the login process chown the these to to the
current user (and something seems to restore them during or after
logout).
I don't know enough about the internals of this process.
I did do a couple of experiments with the 'script' command
and 'strace' using commands like:
... and eyeballing the trace file. This shows how the
script command (which uses a psuedo tty -- or pty) searches
for an available device.
I then did a simple 'chown 600 /dev/ttyp*' as root
(this leaves a bunch of /dev/ttyq* and /dev/ttyr nodes
available). The 'script' command then reports that
the system is "out of pty's."
Obviously the script command on my system don't
do a very thorough search for pty's. It effectively
only looks at the first page of them.
The next test I ran was to add a new line to my
/etc/services file (which I called stracetel) -- and
a new line to me /etc/inetd.conf that referred to it.
This line looks like this:
... all on one line, of course.
Then I connected to that with the command:
This gives me an strace of how telnetd handles the
allocation and preparation of a pty. Here, as I suspected,
I saw chown() and chmod() calls after telnetd did it's
search through to list of pty's to find the first one.
Basically both programs (and probably most other
pty clients) attempt to open each pty until one returns
a valid file descriptor or handle. (It might be nice
if there was a system call or a daemon that would allow
programs to just say "give me a pty" -- rather than forcing
a flurry of failed open attempts -- but that's probably too
much to ask for.
There result of these experiments suggests that there
are many ways of handling pty's -- and some of them may
have to be set as compile time options for your system.
It may be that you just need to make all the pty's
mode 666 (which they are on my system) or you might
chgrp them to a group like tty or pty, make them mode
660 and make all the pty using programs on your system
SGID.
I've noticed that all of my pty's are 666 root.root
(my tty's root.tty and ttyS*'s are root.uucp all are
mode 660 and all programs that need to open them are
either root run (getty) or SGID as appropriate).
Some of the policies for ownership and permissions are
set my your distribution. Red Hat 2.x is *old* and
some of these policies may have changed in the 3.03 and
4.1 releases. Mine is a 3.03 with *lots* of patches,
updated RPM's and manually installed tarballs.
Frankly I don't know *all* of the security implications
of having your /dev/tty* set to mode 666. Obviously
normal attempt to open any of these while they're in
use return errors (due to the kernel locking mechanisms).
Other attempts to access them (through shell redirection,
for example) seem to block on I/O. I suspect that a
program that improperly opened it's tty (failed to
set the "exclusive" flag on the open call) would be
vulnerable.
Since you're an emacs fan -- maybe you can tell me --
is there an mh-e/Gnus IMAP client?
No Kyle Jones (VM maintainer/author) has said maybe IMAP4 for VM
version 7. I think his idea is to make VM do it what it does well and
rely on outside packages to get the mail to it ...
Also -- isn't there a new release of ange-ftp --
I forget the name -- but I'm sure it changed named too.
Yes it's called EFS - it preserves all the functionality but is more
tightly meshed with dired - supposedly it will be easier to use EFS in
other elisp packages (I don't know why or how this would be so).
I'll have to play with those a bit.
Can VM handle mh style folders?
--
Jim
From: David J. Weis, weisd3458@uni.edu
Glancing ahead -- I'd guess that this would take quite a bit
more than a few minutes.
My company has a domain name registered (plconline.com) and two offices.
One is the branch office which is located in the city with the ISP. The
head office is kind of in the sticks in western Iowa. I've been
commissioned to find out how difficult it would be to set up the uucp so
the machine in Des Moines (the big city ;-) would grab all the domain mail
and then possibly make a subdomain like logan.plconline.com for all the
people in the main office to use email.
This would all be running on RedHat 4 over dialup uucp. The system in Des
Moines uses uucp over tcp because it has to share the line with
masquerading, etc.
Thanks for any advice or pointers you have.
My uucp mostly works but I haven't configured it to
run over TCP yet. I also haven't configured my
system to route to any uucp hosts within my domain.
You can address mail to a uucp host through a
DNS by using the '%' operator. For example I can
get my main mail system (antares.starshine.org) to
forward mail to my laptop using an address like:
... the DNS MX record for starshine.org routes
mail to my ISP. My ISP then spools it up in UUCP
until my machine (antares) picks it up. The
name antares is basically transparent to most of
this process.
When antares gets the mail it converts the
percent sign into a "bang" (!) and spools it
for mercury (which happens to be my laptop).
Obviously requiring all of your customers and
correspondents to use percent signs in their addressing
to your users is not going to work very well. It will
probably result in alot of lost mail, alot of complaints
and a constant barrage of support calls.
There are two ways to make your internal mail routing
transparent to the rest of world. You can create a
master aliases list on your mail hub (the easy way) or
you can create DNS and MX entries for each of the hosts.
If you'd like more help we could arrange to talk on
the phone. UUCP is difficult to set up for the first
time (nearly vertical initial learning curve). Once it's
set up it seems to be pretty low maintenance. However
my meta-carpus can't handle explaining the whole process
via e-mail (and I don't understand enough of it well to
be brief).
--
Jim
From: Barry, remenyi@hotmailcom I run Redhat 4.1 with mtools already installed, with it, I can copy a
file to or from a dos disk in A: with mcopy etc..
But if I change the disk & do mdir, it tells gives me the listing of
what was in the last disk. The only solution is to wait hours for the
cache to expire before I can look at another disk.
The problem occurs no matter how I access the floppy, I also tried using
dosemu, and mount, but I have the same problem. I can read and write
from the first disk that I put in with no problems, but if I change the
disk, the computer acts as if the first disk is still in the drive. It
also doesn't matter who I am loged in as eg. root has the same problem.
I also upgraded mtools to 3.3 but no change.
Is there some way to disable the disk cache (I assume thats the problem)
for the floppy drive?
You probably have a problem with the "change disk" detection
circuitry on your floppy.
There's a pretty good chance that you'd see the same thing
under DOS too.
Unfortunately I don't know of an easy way to solve this
problem. You could try replacing the floppy ($30 or so)
the controller ($20 -- to ???) and/or the cable.
If that's not feasible in your case you could try something
like a mount/sync/umount (on a temporary mount point).
This might force the system to detect the new floppy. It's
very important not to try to write anything to a floppy when the
system is confused about which floppy is in there.
DOS systems that I have used -- while they were afflicted
with this problem -- sometimes severely trash the directories
on a diskette in that situation.
It probably doesn't even matter if the mount, sync, umount
that I describe fails -- just so the system is forced to
"rethink" what's there. I'd consider writing a short script
to do this -- put a temporary mount point that's "user" accessible
to avoid having to be root to do this (and especially to avoid
having to create any SUID root perl scripts or write a C wrapper
or any of that jazz).
Here's a sample line for your /etc/fstab:
(according to my man pages the "user" options should
imply the nosuid, nodev etc. options -- which prevent
certain other security problems).
So your chdisk script might look something like:
... you could also just do a 'mount /mnt/tmp' or a
'mount /mnt/a' or whatever you like for your system --
and just use normal Linux commands to work with those
files. The mtools are handy sometimes -- but far from
indispensable on a Linux system with a good fstab
file.
As a security note: mount must be SUID in order to
allow non-root users to mount filesystems. Since
there have been security exploits posted on mount
specifically and various other SUID files chronically,
I suggest configuring mount and umount such that they
can only be executed by members of a specific group
(like a group called "disk" or "floppy"). Then you
can add yourself and any other users who have a valid
reason to work at your console to that group. Finally
change the permissions on mount and umount to something
like:
.... i.e. don't allow "other" to execute it.
This also applies to all your SVGALib programs (which
should not be executed except from the console) and
as many of your other SUID programs as you can.
(... it would be nice to do that to sendmail -- and
I've heard it's possible. However it's a bit trickier
than I've had time to mess with on this system).
As PAM (pluggable authentication module) technology
matures you'll be able to configure your system to
dynamically assign group membership's based on
time of day and source of login (value of `tty`).
This will be nice -- but it doesn't appear to be
quit ready yet.
--
Jim
I just wanted to write to thank you for you response to my mail.
I did as you suggested and the problem is solved!
Actually, you were also right about the problem occurring in DOS as
I used to have a lot of floppies go bad before I went all the way
to linux, but I didn't make the connection.
Anyway, thanks again, you've made my day!
Barry
You're welcome. I'm glad it wasn't something complicated.
BTW: which suggestion worked for you? Replacing one or
another componenent? Or did you just use the "mount, sync,
umount" trick?
Under DOS I used to use Ctrl-C, from the COMMAND.COM A:
prompt to force disk change detection. You can use that
if you still boot this machine under DOS for some work.
--
Jim
From: Benjamin Peikes, benp@npsa.com 1) I'm using one machine with IPAliasing and was wondering if
there is a version of inetd built so that you can have different
servers spawned depending on the ip number connected to.
That's an excellent question.
There is apparently no such feature or enhanced version of
inetd or xinetd.
It also doesn't appear to be possibly to use TCP Wrapper
rules (tcpd, and the /etc/hosts.allow and /etc/hosts.deny)
to implement this sort of virtual hosting.
So far it appears that all of the support for virtual hosting
is being done by specific applications. Apache and some other
web servers have support for it. The wu-ftpd's most recent
versions support it.
I suspect that you could create a special version of
inetd.conf to open sockets on specific local IP addresses
and listen on those. I would implement that as a command
line option -- passing it a regex and/or list of ip addresses
to listen on after the existing command line option to
specify which configuration file to use. Then you'd load
different copies of this indetd with commands like:
(This would be something like -- all of the 192.168.14.*
address and all of the 17.18.*.* addresses are handled by
the first inetd -- all of the access to a host named
barneyweb (presumably looked up through the /etc/hosts file)
would be handled by the next inetd. and all of the accesses
to the ipalias 192.168.2.3 would be handled by the last one)
This would allow one to retain the exact format of the
existing inetd files.
However I don't know enough about sockets programming to
know how much code this would entail. The output of
'netstat -a' on my machine here shows the system listening
on *:smtp and *:telnet (among others). I suspect that those
stars would show up different if I had a socket open to
a specific service on a specific service.
This scheme might use up to many file descriptors. Another
approach would be to have a modified tcpd. This would have
to have some option where by the destination *as well as*
the source was matched in the /etc/tcpd.conf file(s).
(Personally I think that tcpd should be compiled
with a change -- so that the single tcpd.conf
file is used in preference to the /etc/hosts.allow
and /etc/hosts.deny files. Current versions do
support the single conf file -- but the naming is
still screwy).
I'm not sure quite how Wietse would respond to this --
possibly by repeating the question:
"If you want me to add that -- what should I
take OUT?"
(which is what he said one to me when I suggested merging
his chrootuid code with tcpd).
I've blind copied Wietse on this (Hi!). I doubt he has
time to read the Linux Gazette.
2) A related problem: I have one machine running as a mail server
for several domains where the users are using pop to get their
mail. The problem is that the From: line always has the name
of the server on it. Is there a way to use IPaliasing to fix
this? Or do I have to muck around with the sendmail.conf file?
This is becoming a common question.
Here's a couple of pointers to web sites and FAQ or HOWTO
documents that deal specifically with "Virtual Mail Hosting"
(look for references to "virtualdomains")
... and here's one guide to Virtual Web Hosting:
I guess the best way to do this would be to change inetd to figure
out on which interface the connection has been made on and then
pick the correct inetd.conf to reference, like
I would recommend that as a default behavior.
I suggested adding additional parameters to the
command line specifically because it could be done
without breaking any backward compatibility. The
default would be to simply work as it does now.
I still suspect that this has some scalability problems
-- it might not be able to handle several hundred or several
thousand aliased addresses.
I might still be useful to implement it as a variation of --
or enhancement to -- tcpd (TCP_Wrappers).
I think that inetd reads in the configuration file when it
starts because it needs a SIGHUP to force it to reread the conf
file. All you would have to do is make it reference the right table.
This is also documented in the inetd man page.
Do you know where I could find the code? I would be interested
in looking at it?
The source code from inetd should be in the bundle
of sources that comes with the "NetKit"
Look to:
ftp:..ftp.inka.de/pub/comp/Linux/networking/NetTools/
and mirrored at:
ftp://sunsite.unc.edu/pub/Linux/system/network/NET-3-HOWTO/
... this includes the history of it's development and the
names of people who were active in it at various stages.
If you're going to try to hack this together -- I'd suggest
a friendly posting to the comp.linux.development.system
newsgroup -- and possibly some e-mail to a couple of
carefully chosen people in the NET-3-HOWTO.
--
Jim
From: John Doe
Well, here's someone who wants to make a anonymous
tip to "The Answer Guy."
At "John Doe's" request I looked over this site. It
does have extensive information about modems -- including
lots of press releases about which companies are acquiring
each other (3Com over US Robotics, Quarterdeck gets DataStorm).
However there didn't appear to be any references to Linux,
Unix or FreeBSD.
So -- if one needs information about modems in general this
looks like an excellent site to visit. However it the question
pertains specifically to using your modem with Linux -- I'd
suggest:
http://sunsite.unc.edu/LDP/HOWTO/Serial-HOWTO.html
--
Jim
From: Yang, lftian@ms.fudan..edu.cn
I have an AT 3300 card( from Aztech) which integrates the function of
sound card and 28.8K modem. It seems that it need a special driver for its
modem function to be work. In MSDOS, there is a aztpnp.exe for that
purpose. Do you know is there any way I can get the card work (at least its
modem function) in Linux?
Tianming Yang
I'm not familiar with that device. The
name of the driver suggests that this is a
Plug 'n Play (pnp) device (sometimes we use the
phrase "plug and *pray*" -- as it can be a toss
of the dice to see if they'll work as intended.
My guess would be that this is a PCMCIA card
for a laptop system (which I personally pronounce
"piecemeal").
Did you look in the "Hardware HOWTO" (start at
www.ssc.com, online mirror of FAQ's and HOWTO's)?
Did you go to Yahoo! and do a keyword search on
the string:
... (the plus sign is important there)?
Since all of the real details about the configuration
of the card are determined by the manufacturer
(Aztech in this case) I would start by contacting
them.
If they've never heard of Linux -- or express no
interest in supporting it -- please consider letting
them know that Linux support affects your purchasing
decisions. Also let them know that getting support
for Linux is likely to cost them very little.
How to get a Linux driver for your hardware:
If you are a hardware company that would like
to provide support for Linux and FreeBSD and other
operating systems -- but you don't have the
development budget -- just ask.
That's right. Go to the comp.os.linux.development.system
newsgroups and explain that you'd like to provide
full documentation and a couple of units of your hardware
to a team of Linux programmers in exchange for a freely
distributable driver. Be sure to make the sources for
one of your other drivers (preferably any UNIX, DOS, or
OS/2 driver) available to them.
If you don't like that approach, consider publishing the
sources to your existing drivers. If you are really in
the hardware business than the benefits of diverse OS
support should far outweigh any marginal "edge" you might
get from not letting anyone see "how you do it."
(Just a suggestion for all those hardware vendors out there).
--
Jim
From: Dani Fricker, 101550.3160@CompuServe.COM
The IP Address of any connecting client is provided
to any CGI scripts you run, and is stored in the
server's access log (or a reverse DNS lookup of it
is stored therein -- depending on your httpd and
configuration).
* Note: I suggest disabling reverse DNS
lookup on webserver wherever possible.
it generates alot of unnecessary traffic
and you can isolate, sort, and look up the
IP addresses in batches when you want to
generate statistics involving domain names.
(I also tend to think that most of the
reports done on web traffic logs have about
as much rigor and resemblance to statistical
analysis as reading chicken entrails).
my ip-gateway connects my inner lan over two token ring network cards
(sorry, not my idea!) with the internet (lan <-> tr0 <-> tr1 <->
internet). the masquerading forward roule of ipfwadm gives me the
possibility to indicate a source and a destination address.
Oh. So all of the clients that you're interested in
are on a private LAN and going through a masquerading/NAT
server (network address translation).
I would try using ident for starters. Run identd on your
Masquerade Host and make calls to the ident service from
your CGI scripts. I don't think it will work -- but it
should be worth a little info.
From there you might be able to configure all the clients
on the inner LAN to use an *applications* level proxy
(squid -- formerly cached, CERN httpd, or the apache cache/
proxy server). Masquerading can be thought of as a
"network layer proxying services" while SOCKS, and similar
services -- which work with the co-operation of the client
software -- are applications layer proxies.
I don't know if the private net IP address or other info
will propagate through any of these HTTP proxies.
If this is *really* important to you, you could consider
writing your own "NAT Ident" service and client. I don't
know how difficult that would be -- but it seems like the
code for the identd (and the RFC 931? spec) might give you
a starting point for defining a protocol (you might want
to secure that service under TCP_Wrappers). You might want
to consider making this a TCP "Multiplexed" service --
look for info on tcpmux for details about that.
The gist of tcpmux is that it allows your custom client
to talk to a daemon on TCP port 1 of the server host and
ask for a service by name (rather than relying on
"Well-Known Port Addresses"). So, if you're going to create
a new service -- it makes sense to put it under tcpmux
so you don't pick your own port number for it -- and then
have the IANA assign that port to something else that you
might want later.
do you see a possibility for an 'address assignment' between the two
interfaces? if you do please let me know.
I don't know of any existing way to determine the IP
address of a client on the other side of any NAT/masquerading
host -- I'm not even sure if there's any existing way to do it
for a client behind a SOCKS or TIS FWTK or other applications
level proxy.
I'll be honest. With most "Answer Guy" questions I
do some Yahoo!, Alta-vista and SavvySearch queries -- and
ask around a bit (unless I already know the answer pretty
well -- which doesn't happen all that often these days).
I skipped that this time -- since I'm pretty sure that
there's nothing out there that does this.
I welcome any corrections on this point. I'll be happy
to forward any refutations and corrections to Dani.
All of this begs the greater question:
What are you really trying to do?
If you are trying to provide some form of transparent
access control to your webserver (so local users can
see special stuff without using a "name and password")
-- there are better ways available.
Netscape and Internet Explorer both support a form
of client-certificate SSL -- which is supported at
the server side by the Stronghold (commercial Apache)
server.
As an alternative -- I'd look at the possibility of
finding or writing a Kerberos "auth" module for
Apache (and deploying Kerberos to the clients).
This might be more involved than you're management
is willing to go for -- but writing new variations of
the indentd service might also fall into that category.
IP addresses are a notoriously bad form of access
control. If you have a properly configured set of
anti-spoofing rules in the packet filters on your
router -- and you can show that no other routes exist
into your LAN -- then you can base access controls to
services (TCP/Wrappers) to about the granularity of
"from here" and "not from here." Attempting to read
more into them than that is foolhardy.
Ethernet and Token Ring MAC (media access control) addresses
(sometimes erroneously called "BIA's" -- burned in addresses)
are just about as bad (most cards these days have options to
over-ride the BIA with another MAC -- usually a feature of
operating the card in "promiscuous" mode).
Yet another approach to the problem might be to simply
put a web server on the internal LAN (no routing through
the NAT/masquerading host) -- and use something like
rdist to replication/mirror the content between the
appropriate document trees on the internal and exterior
web servers.
Basically we'd need to know much more about your
requirements in order to give relevant recommendations.
--
Jim
From: Mohammad A. Rezaei, rezaei@tristan.TN.CORNELL.EDU
I absolutely agree. I wonder where I suggested 'dd'
without expressing my misgivings.
Please consider quoting little portions of my posting
when making references to them -- I write alot and
can't remember past postings without some context.
I have more than once installed/transfered entire hard drives using
tar. simply put both drives in the same machine, mount the new drive
in /mnt and do something like
There are better ways to do this.
One way is to use a command like:
In both of these cases you can use find parameters
to include just the files that you want. (Note:
with tar you *must* prevent find from printing any
directory names by using the -type f (or more
precisely a \! -type d clause) -- since tar will
default to tar'ing any directories named in a
recursive fashion).
The -T (capital "tee") option to GNU tar means to
"Take" a list of files as an "include" list. It
is the complement to the -X option that you list.
You can also pipe the output of your find through
grep -v (or egrep -v) to filter out a list of
files that you want to exclude.
finally, one has to install the drive onto the new machine,
boot from floppy and run lilo.
The disks don't have to be identical. the only disadvantage is having
to run lilo, but that's takes just a few minutes.
The only message I can remember posting about 'dd'
had an extensive discussion of using tar and cpio for
copying trees.
Am I forgetting one -- or did you only get part of
my message?
Hope this helps.
Hopefully it will help some readers. The issues of
copying file trees and doing differential and
incremental backups is one that is not well covered in
current books on system administration.
When I do a full backup I like to verify that it
was successful by extracting a table of contents or
file listing from the backup media. I then keep a
compressed copy of this. Here I use tar:
.... where the contents list is named something like:
.... which is a hostname, a volume (tape) number and a
date in YYYYMMDD format (for proper collation -- sorting).
To do a differential I use something like:
... (actually it's more complicated than that since
I build the list and compute the size -- and do some
stuff to make sure that the right volume is on the
Magneto Optical drive -- and mail nastygrams to myself
if the differential won't fit on that volume -- if the
volume is the most recent one (I don't overwrite the
most recent -- I rotate through about three generations)
-- etc).
However this is the core of a differential backup.
If you wanted an incremental -- you'd supply a different
file to the -newer switch on your find command.
The difference between differential and incremental is
difficult to explain briefly (I spent about a year
explaining it to customers of the Norton Backup). Think of
it this way:
If you have a full -- you can just restore that.
If you have a full, and a series of differentials,
you can restore the most recent full, and the
most recent differential (any older fulls or differentials
are unneeded)
If you have a full and a series of incrementals you
need to restore the most recent full, and each
subsequent incremental -- in order until the most
recent.
It's possible (even sensible in some cases) to use a
hybrid of all three methods. Let's say you have a large
server that takes all day and a rack full of tapes to do a
full backup. You might be able to do differentials for
a week or two on a single tape per night. When that fills
up you might do an incremental, and then go back to
differentials. Doing this to a maximum of three incrementals
might keep your all day backup marathons down to once a month.
The restore must go through the "hierarchy" of media in the
correct order -- most recent full, each subsequent incremental
in order, and finally the most recent differential that was
done after that.
(Personally, I avoid such complicated arrangements like the
plague. However they are necessary in some sites.)
-- Jim
Then open the file you just created with a text editor, and change any
information that applies to the secondary ISP, eg. dialup, the IP number
of the ISP, username and password. write the file(save it) and try your
new executable, ppp-on.anysuffix. Just a quick pointer, you could call
your new script any name you want as long as there's no other file with
the same name in your path,preferably no other file with the same name at
all
These days most Internet Service Providers assign you a Dynamic IP when
you logon to their network, due to the cost of assigning every customer
a static IP. At present there are only so many IP addresses available
and, apparently each one costs to register. Consequently ISPs buy a pool
of IP addresses within a range and assign an available one at login. For
most uses, such an arrangement is no problem, assuming that most
internet usage consists of interaction between the ISP's computer and
the local one. For some purposes, however such as allowing telnet or ftp
to your computer the dynamic scheme is less than ideal. Here's
a relatively painless way to get your current IP, so you can run with
the big dogs. Open an xterm, or rxvt and type:
which will bring up some info in two blocks. You'll want to note the
bottom block, which will have a line that specifies your inet address
expressed numerically. It will be in a xxx.xxx.xxx.xxx format, which
corresponds to the standard IP address, in fact that's what it is.
you ca write this number down or just highlight this IP address(to paste
it)and type:
the number being your inet address from the last step.
It may take a couple of minutes, but you will get a two line messagethat
looks like:
we may talk about FQDN some more another time, but for the purpose at
hand, just type:
where Name is the first line from the above step. That's it, except that
you must repeat this procedure every time you connect to your ISP. You
might be able to write a script to automate this procedure, but in the
meantime which as my friend Al used to say "is a groovy time", you can
use this knowledge to run remote X apps(just a minute,I'm coming to that)
allow your friends or inet associates to telnet to your computer, or ftp
files from a telnetted site(this too, momentarily).What you need to know
is that the next time youreboot, you may get a message saying that your
computer name is "bad". This isn't a comment on your lack of originality
or taste, and you should basically ignore it.
as detailed above. If you have a static IP address, you can skip this
step.What needs to be done next is to type:
When you hit enter you will see a message like, "the.telnet.box has been
added to the control list". You will probably have to restart your window
manager, your mileage may, as they say, vary. Now when you start a
telnet session, you can enter the name of an X application and in a
moment, the application window will apear on your screen, even if you
don't have it installed on your computer. Do your work, play your game,
and marvel at the ramifications of this capability.
You can also invite friends and coworkers over to your computer to
do some work, socialize or learn something, in the following manner.
Obtain your FQDN, or IP address, as detailed above. E-mail it to them or
call them on the phone to let them know where you are today(Not where
you want to go today, that's another "OS"). they can then:
and all of a sudden they're in your den, or office or wherever you keep
your computer. For more sophisticated methods of getting your address,
read the "Dynamic IP hacks-HOWTO.
This esteemed tabloid is just full of novice- to intermediate level
tips and tricks.The Answer Guy, Two Cent Tips, and The Weekend Mechanic
in particular, are good sources of the kinds of things that will make you
a demi-guru in no time at all.
In DOS and Windows,formatting a floppy disk is a one shot affair
which formats, erases data, and creates a file system on the floppy.
In linux however, you have to format and create the filesystem in
separate steps. At first glance, this seems backward, after all, isn't
linux a more sophisticated OS? Why do things in two steps that the others
do in one? The reason is that linux can read several filesystems so that
data can be moved from one OS to the other. By mounting the floppy drive
as MSDOS, VFAT, or other filesystem type the data can be read from the
mount point in a manner that linux can make use of.
have trouble with the
command?Try leaving your computer on overnight, and
the next day use the command instead.
Locate is a database that is gathered during idle times on your computer
Actually locate reads a database that is updated by a command in your
system files, but if you give it time to breathe, usually overnight, it
can locate any file on your hard drive(s). You can also try
and you will get a location for the named file.
TTYL, Mike List
In this article I will describe a configuration tool called The Dotfile
Generator (TDG for short). TDG is a configuration tool, which configures
programs, using X11 widgets like check boxes, entries, pull-down menus etc.
For TDG to configure a given program a module must be made for it. At the
moment modules exist for the following programs: Bash, Fvwm1, Fvwm2, Emacs,
Tcsh, Rtin and Elm.
The article will describe common use of TDG, so if you do not have it yet,
it might be a good idea to download
it (It's free!) You may also go to the home page of the
Dotfile Generator for further information.
A basic concept in UNIX is that the programs are very configurable. Here is
an example from Emacs, which shows this:
This solution, however, requires that the user has to learn the programming
language used in the dot-file, and has to read lots of documentation to
find out which configurations can be made. This task may be difficult and
tedious, and for that reason many users often choose to use the default
configuration of the program.
If you take a look at some dot-files, you may find
that most of the configurations can be described by the following items:
When you start TDG, you will be offered a list of standard configurations,
where you may pick one to start out with. This may be convenient, if you do
not have a dot-file for the given program, or if you would like to try a
new configuration. If on the other hand, you already have a dot-file, which
you would like to put the finishing touches to, you may read this file into
TDG. Note, however, that it is not all modules, which have the capability
to read the dot-file (the fvwm2, rtin and elm modules have, the other
modules do not, since it would be to complicated to create such a
parser.)
When you have selected a start-up configuration, the menu-window will be
displayed (see figure 1). In this window, you can travel through the
configuration pages, just like a directory structure. If you select a
page, a new window will be displayed, with the configuration for this page
(see figure 2). This window will be reused for all the configuration pages,
ie. only one configuration page is visible at a time, so you do not have
to destroy the window yourself.
Figure 1 Figure 2
In region 3, information is shown on what will be generated. You have
three possibilities:
When you have done all the configurations, you have to tell TDG which file
you wish to generate. This is done from the Options menu
(Setup->Options). And now it's time to create the actual
dot-file, which is done by selecting Generate in the File
menu.
Once you have generated the dot-file, you may find that you would like some
of the configuration to be different. You could now go to the configuration
page in question, change your configuration, and then generate once again. If,
however, you are testing several different options for a single
configuration (ie. several items from a pull-down menu) you may find it
cumbersome to generate the whole module over and over again. In this
situation, you may chose Regenerate this page in the File
menu. Note, however, that if some part of the configurations on the page effects
other pages, these will not be generated, so in these situation
you have to generate the whole module.
To see how to use the generated dot-file, please go to the Help
menu, and select the How to use the output item.
This ExtEntry has three visible tuples, though only two of them contain
values (you can see, that the third one is grayed out). To add a new tuple
to the ExtEntry, you have to press the button in the lower right corner,
just below the scroll bar. If the ExtEntry contains more tuples than can be
shown in it, you may scroll to the other tuples with the scroll bar.
If you press the left mouse button on one of the scissors, a menu
with four elements will be displayed. These elements are used to cut, copy
and paste tuples within the ExtEntry.
If the tuples get very large, only one of them may be shown on the screen
at a time. An example of that is seen in figure 4.
When the tuples contain many widgets, scrolling the ExtEntry becomes
slow. In these cases, the ExtEntry may have a quick index. In figure 4, you
can see the quick index at the top of the ExtEntry (it's the button labeled
Idx.) When this quick index is invoked, a pull-down menu is display with
the values of the element associated with the quick index. This makes it much
easier to scroll the ExtEntries.
Figure 4 Figure 5
In TDG, a special widget has been created called a FillOut, which
does configurations like the above. In Figure 5, you can see a FillOut
widget from the Bash module. At the top of the widget there is an entry,
where you can type ordinary text. Below it, the tokens are placed. If you
select one of the tokens, it is inserted in the entry at the point of the
cursor. Some of the tokens may even have some additional
configurations. Eg. the token Current working directory has two
possible options: Full directory, and only the last part. When tokens with
additional configurations are selected, a window will be displayed, where
these configurations can be done. If you wish to change such a configuration,
press the left mouse button on the token in the entry.
The widgets will appear as a button within TDG, and when the button is
pressed a new window will be displayed, where the actual configuration is done.
Next time you enter TDG, your saved file will be one of the the files you
will be offered as a start-up configuration.
One important point you have to note is that this save file is an
internal dump of the state of TDG. This means that this file dependson
the version of TDG and the module. This means that if you wish to send a
given configuration to another person, this format is not appropriate. A
version independent format exists, which is called the export
format. To create such a file, you have to select Export
instead of Save in the File menu.
Sometimes you may wish to restore the configuration on a single page, to its
value as it was before you started playing around with it, or you may wish
to merge another person's configuration with your own. This is done by selecting
Reload in the File menu. To tell TDG that you only want
to reload some of the pages, you have to select the Detail button
in the load window. This will bring up a window, where you can select which
configuration pages, you wish to reload. Here you can also tell it how you
want the pages to be reloaded. You have two possibilities:
It's always a good idea to have a bookmark on this page, as work is
currently in progress on new modules.
It's not quite 'ready for prime-time', meaning there is
almost no documentation
and there is still a lot of work to do on the user interface.
http://www.cs.kuleuven.ac.be/
ELECTROGIG Technology
For more information, check out
http://www.ma.utexas.edu/~mzou/EZWGL.
libgr will build shared libs on Linux-ELF and on HP/UX.
EPSCAN can be found at
The driver should work for any of the
ES-{300-800}C / GT-{1000-6500} models as well, but has not been
tested on these.
SIMLIB IG for Linux is $2500 (US)
KNIENIEDER Simulationstechnik KG
(office@knienieder.co.at)
...that there is a freely available RenderMan shader library from
Guido Quaroni? The library contains shaders from the RenderMan
Companion, Pixar, Larry Gritz and a number of other places.
You can find a link to it from the BMRT Web pages at
http://www.seas.gwu.edu/student/gritz/bmrt.html.
...that there is an FTP site at CalTech that contains a large
number of RenderMan shaders? The collection is similar to Guido
Quaroni's archive, except the FTP site includes sample RIB files
that use the shaders plus the rendered RIBs in both GIF and TIFF
formats. The site is located at
ftp://pete.cs.caltech.edu/pub/RMR/Shaders/.
Q and A
Q: Where can I get a copy of the netscape color cube for
use with Netpbm? How should it be used?
A: The color cube can be found at the web site for
the text Creating Killer Websites at
http://www.killersites.com/images/color_cube_colors.gif.
The cube can be used in the following manner:
Q: Where can I get models of the human figure?
A:
Here are two addresses for human figure models. The first is 3DCafe's
official disclaimer and the second takes you straight to the human
figures. Please read the disclaimer first (although you may need
an asp capable browser, such as Netscape 3.x to do so):
http://www.3dcafe.com/meshes.htm
From the IRTC-L mailing list
Q: Is there a VRML 2.0 compliant browser available for Linux?
A: Yes. Dimension X's Liquid Reality is a fully compliant
VRML 2.0 browser. The download web page says that there will
be support as a plug-in for Netscape 3.x soon. This is a
commercial product with a free trial version available for
download. See
http://www.dimensionx.com/products/lr/download/ for more
details.
From a friendly reader, whose name I
absent mindedly discarded before recording it. My
apologies.
Q: Can anyone tell me how I would go about defining a height field
according to a specific set of data points?
My goal is to be able to take a topographic map, overlay it with a
rough grid, and use the latitude, longitude, and elevation markings
as points in a definable 3-D space to create a height field roughly
equal to real topography.
A: The easiest way is probably to write a PGM file.
I wouldn't use longitude and latidude because the length of one degree
isn't fixed and it will give reasonable results only near the equator.
Use UTM coordinates or superimpose any arbitrary grid on your map
which represents approximate squares.
From Florian Hars via the
IRTC-L mailing list
Q: I've been fiddling with some simple CSG using BMRT and have run
into a problem. I'm trying to cut a square out of a plane that
was created from a simple bilinear patch. Whatever I use to define the
square (a box actually) comes out white instead of the background color
(black in this case). I dont know what I'm doing wrong and was
wondering if someone might take a peek at this for me.
A:
There are several problems with your RIB file, as well as your use
of CSG. The two biggies are:
You just can't do this:
From Larry Gritz <lg@pixar.com>
Linux Graphics mini-Howto
Some of the Mailing Lists and Newsgroups I keep an eye on and where I get alot
of the information in this column:
The Gimp User and Gimp Developer Mailing Lists.
DESCRIPTION OF HOW THIS IMAGE WAS CREATED:
Light Source Shaders
Surface Shaders
Volume Shaders
Displacement Shaders
Transformation and Imager Shaders
Shader procedure names
Variables and scope
Data types and expressions
Functions
Note: This particular example might not be very useful. It is just meant to
show how to include functions from a function library.
Functions are only callable by the shader, not directly by the renderer.
This means a function cannot be used directly in a RIB file or referenced
using the C binding to the RenderMan Interface. Functions cannot be
recursive - they cannot call themselves. Also, all variables passed to
functions are passed by reference, not by value. It is important to
remember this last item so that your function doesn't inadvertantly make
changes to variables you were not expecting.
Statements
Coordinate Systems
In shaders, [we] break down complicated surface patterns and textures into
layers. Each layer should be fairly easy to write (if not, then we can
break the layer into sub-layers). Then, [we] combine the layers by
compositing.
A colored cross pattern
Adding opacity - a wireframe shader
A simple paper shader
A textured mapped chalkboard
Displacement map example
Next month I planned on doing the 3rd part of this 3 part BMRT series. I
think taking 2 months between articles worked well for me this time since
it allowed me a little more time to dig deeper. Plan on the final article
on BMRT in this series in the July issue of the Graphics Muse. Till then,
happy rendering.
Kandinski is my new pre-pre-pre-beta program which generates a picture
file from a MIDI file. It does so based on my cycluphonic method of
correlating colors to musical pitches. The few careful observers who have
seen previous implementations of cycluphonics agree that it gives visual
events which seem to sympathize with the generating music, in terms of
implied feeling, better than previous "color organ" methods.
Kandinski was written with pfe under Linux on a 486. It should be easy to
port to another ANSI Forth system, as I am rusty at Forth, and the task at
hand didn't call for any trickery, and I avoided the Linux-specific stuff
in pfe, mostly because I couldn't find much documentation on it.
The code presented here creates a .ppm image file on a selectable track by
track basis. The piano envelope option is not implemented yet, just organ.
.ppm files can be converted to just about any image format with the unix
pbmplus tools, and are viewable in Linux with zgv.
The crucial cycluphonic element in Kandinski is the "cycle" construct,
a lookup table which Kandinski uses to map a 12 hue color wheel to the
Cycle of Fifths. That's the crux of cycluphonics. If you use this code, or
cycluphonics, give credit where due.
Separate documentation file for the Kandinski program
Rick Hohensee http://cqi.com/~humbubba
or rickh@capaccess.org
please cc to humbubba@cqi.com
"Well, should we get one pitcher or two?" That was the question
that began the first unofficial event of the Linux Expo Thursday
night. A group of people, including Red Hat employees, some of the
speakers and a tired maddog were at the Carolina Brewery in
Chapel Hill, North Carolina. It was late, and I was the
last person to arrive."Two pitchers," I cried,"Now what
will you be drinking?"
The next day, Friday April 4th started early, as I had to set up
the Linux International booth, as well as absorb all that was
happening. The event was held in the North Carolina
Biotechnology Center at Research Triangle Park. As I
approached the Biotech Center, I was met with a friendly
parking coordinator that reinforced the information that
"parking was scarce", and that most people had to park at
outlying lots. Fortunately Red Hat had arranged for shuttle
busses from those lots and from several of the hotels. Since our
car had an exhibitor's pass, we were able to park close to the
Biotech Center and unload our banners, handouts and stuffed
penguins.
There was a large tent to the left outside of the building
containing the "Linux Expo Super Store" stocked with
Linux books, Linux bumper stickers, T-shirts (including
an excellently designed Expo shirt that said "Expose yourself to
Linux" with a front and rear view of a penguin holding open an
overcoat) and other interesting souvenir items. Further
to the left was an outdoor viewing area for the conference talks
that (due to the excellent weather) was a favorite spot for
people to view the technical talks for free, especially while
playing Frisbee. A raffle was held in the registration area,
and prizes were given away on an hourly basis.
Having registered, attendees were given a copy of the talks as
well as an event schedule.
The event was held on two floors with the exhibits spread out on both. There
was another conference viewing area inside the building with TV
monitors, as well as the conference auditorium itself. There was
an Install Fest area (sponsored by the Washington D.C. Linux
User's Group, Linux Hardware Solutions and Red Hat Software),
where people brought their systems, received help with installing
Red Hat's latest release, and Olaf Kirch's kernel-based NFS
server was "stress tested" at the same time. Finally, there was a food
court area, where people could buy sandwiches, chips, soda and
other "software development food".
There were fifteen vendors at the Expo, each with "table-top"
booths to display their wares. I prefer the "pipe-and-drape"
approach to trade shows rather than expensive booths, since I
would rather the vendors put more money into development of the
product and less into elaborate displays or floor shows with
unicycle riders who juggle things. While not all Linux vendors
were at Linux Expo, a wide spectrum of companies, including Linux
International, Cyclades, Numerical Algorithms Group, Linux
Hardware Solutions, Enhanced Software Technologies, Caldera,
Applix, Xess, WorkGroup Solutions, Stay Online, VA Research, Apex
Systems Integration, PromoX Systems and (of course) Red Hat
Software were present. One item being demonstrated at the Linux
Hardware Solutions booth was a free piece of software called [cw]em86[ecw]
that allowed an Intel/Linux binary to run without change on an
Alpha/Linux system. Being shown for the first time, it allowed
Applixware, Netscape and various other applications to execute as
if they had been ported to the system.
Penguins abounded in various T-shirts, giveaways and objects
d'art. In fact, there were so many people there (I estimated 900
over the two-day event) with penguin "stuff", that I thought
I'd
had enough of penguins; but afterwards while wandering around Chapel Hill, Alan Cox found some
candy in the shape of penguins, so penguin "lust" started all
over again.
The technical conference started off with a presentation by
Gilbert Coville of Apple Computer with a talk about the MkLinux
kernel. For people who were afraid that this would turn into a
"Red Hat Only" event, it was interesting that Gilbert's talk
opened the Expo and that a talk about the Debian Linux
Distribution (given by Bruce Perens) followed shortly after.
Bruce also discussed the graphics used in the making of Toy Story in
a separate presentation.
Various presentations about hardware-specific ports were given.
Dave Miller talked about the "Next Generation SPARCLinux" as well
as the Free Software Development Model, and David Mosberger-Tang
talked about the Alpha Port, as well as methods, applicable to both Intel
and Alpha, for
speeding up your programs by paying attention to memory and
cache accesses.
Other talks were more general across the Linux OS, such as Jeff
Uphoff's "Network File Locking", Alan Cox's "Tour of the Linux
Networking Stack", Peter Braam's "Coda Filesystem", Alexander
Yuriev's talk on the IPv4 family of protocols and infrastructure
and his talk on security, Michael Callahan's "Linux and Legacy
LANs", Eric Youngdale's "Beyond ELF", Olaf Kirch's "Linux Network
File System", Theodore Ts's "Ext2 File System: Design,
Implementation and the Future", Miguel de Icaza's talk on the new
RAID code and Daniel Quinlan's talk on the File System Hierarchy
Standard.
To round out the list of talks and events was Dr. Greg
Wettstein's talk on "Working and Playing with others: Linux Grows
Up" and the Linux Bowl.
The Linux Bowl was the final event. Two teams
of six developers were pitted against each other to answer
thirty questions about Linux and the Linux community. Questions
ranged from "What liquid should one drink between rounds of a
Finnish sauna?" (correct answer: beer) to "What version library
fixed a particular security hole?" to which Alan Cox gave a
(seemingly) ten minute answer. While some of the questions were
very obscure (even the moderator was unsure of the answer), most
of the time either the right answer (or a good facsimile) was
given.
The show sponsors (after tallying up the attendence) reported that
958 people showed up, which could be the largest Linux-specific event
ever, of which 40% were from within North Carolina. Attendees came from
over 25 states, 4 Canadian provinces, and 10 countries, including Australia,
Korea and European countries.
Finally, I would like to thank the members of the Atlanta Linux
Enthusiasts http://www.ale.org/ group who helped to staff the Linux
International booth. They were great and helped give me the freedom
to get out from behind the booth every once in a while, because
most importantly, Linux Expo was a chance to talk with the
vendors, the developers and other old and new friends on a
one-to-one, quality basis. Perhaps some things could be improved
for next year: A larger auditorium for the talks, more and
closer parking and less expensive food in the food court. But
certainly the southern hospitality and warmth of Red Hat Software
came through. I want to thank the sponsors for arranging a great event,
and I hope that next year's will be even larger and better.
Most of the window-managers available for Linux these days can trace their
ancestry back to the original twm program, which may have been the first
widely used manager on unix systems. There is a good reason for this, as twm
pioneered many of the features taken for granted by users, such as movable,
resizable windows and a root-window applications menu. It's good, time-tested
code; why reinvent the wheel?
Two programmers have recently done just that, from two perspectives as far
removed from each other as their respective geographical locations. Chris
Cannam, a British programmer, has taken the minimalist approach with his wm2
manager (which I wrote about in LG #14) and the new wm2 variant wmx, which I
discuss elsewhere in this issue.
At the other extreme is the work of a young Australian programmer who likes
to be known as the Rasterman. Imagine asking the programmers responsible for
the games Quake or Duke Nukem 3D to write a window-manager; the result might
bear some resemblance to the fanciful program known as Enlightenment.
I first encountered Enlightenment (what a name! it seems to carry the
implication that we users of fvwm et al are still crawling blindly through the
primordial ooze...) earlier this year, when a binary was available on the web.
I tried it briefly, but at the time I had a 486 machine; it ran slowly for me
and seemed to consume great gobs of memory. Recently the Rasterman
(his real name is Carsten Haitzler) has rewritten the application from scratch,
tightening it up and introducing a new shared lib which handles image loading
and rescaling. The memory consumption has been greatly reduced since the
initial release. At this point (beta release 4) there are no virtual desktops
or root-window menus, but the project looks promising and what there is of it
runs well for me.
Enlightenment uses the ppm image format for both window details and
icons. An elaborate configuration file (called windowstyles) specifies
which image goes where. Each segment of the window border and detailing is a
separate ppm file. I haven't made any attempt to modify the default
configuration. It looks like it would take many hours to write a new one. Carsten
plans on eventually offering configurations which would emulate any of the
other window-managers.
I get the impression from the Enlightenment web-page that the ppm
format is more efficient than others, especially on 16-32 bit displays. I
don't know how valid this is, but the window-manager does seem to do
quite a bit of image handling without consuming great amounts of memory.
This window-manager automatically will load any sort of image format as a
root background image. At startup the appropriate netpbm utility is summoned
to transform the image to the ppm format. Naturally, you need
to have the netpbm graphics utility package installed for this to work.
Here is a screenshot of a window under Enlightenment:
XV (with which I made the screenshot) couldn't figure out where the actual
window border was; can you blame it? I set the root-window background to be
the same color as this HTML-file background as a quick work-around.
The Enlightenment web-site is at
http://www.cse.unsw.edu.au/~s2154962/enlightenment/index.html.
The source for the latest version can be downloaded from the site; the latest
news about the application will also be there.
It will be interesting to see what eventually happens with Enlightenment,
though personally I'm satisfied with the window-managers I currently use. I
just like to see diversity in software for Linux. Fancy new window-borders
might seem to be a trivial matter but it is user-interface features such as
these which can attract new users, especially younger ones. I showed
Enlightenment to my sixteen-year-old son (an avid computer- game player) and
he was impressed. His comment was "It looks like a game interface!".
Another factor is the simple human desire for novelty. Sometimes the same
old interface becomes boring -- you realize you aren't really even seeing it
anymore. A change in background and window-style can be refreshing. People
routinely change room interiors for these same reasons and, come to think of
it, I look at my computer screen quite a bit more than I do the walls!
Keep in mind that the science-fiction Bladerunneresque appearance is just
the default. Enlightenment is a framework and could be configured in a
variety of ways, depending upon taste (and how much time you're willing to
spend!). Luckily (if you have patience), someone will eventually come up with
a configuration which will suit you. or at least be close. Interest seems to
be growing in this window-manager lately (judging by the volume of messages in
the mailing list) and it may yet evolve into a community-supported
window-manager, such as Fvwm2 or Afterstep. It's been released under the Gnu
license, but so far Carsten Haitzler is the sole developer.
Around 1993 Kevin Burfitt, an Australian computer science student, began
developing a computer program which would transform recorded music into moving
colored patterns. Programs such as this had been in use for some time,
typically as an adjunct to a rock concert, i.e. part of the "light show".
This program was originally written for DOS, though before long it began to
acquire a trait common to software in the unix world: a multitude of options
and parameters.
Kevin must be a fan of the early 20th-century horror writer
H.P. Lovecraft. How else to account for the distinctive appellation "Cthugha"
which he gave his program? In the Lovecraft stories Cthugha is the name given
to a horrific "elder god" which manifested itself to humans in the form of
shifting colored lights. (This doesn't sound too horrific, but Lovecraft
could make a loaf of bread seem sinister!)
Cthugha has from the early days been available under the aegis of
the Gnu General Public License, making the source freely available. This
opened the door for many other programmers scattered throughout the world who
became involved with the project. Sound familiar? Ports of the
program are now available for the PowerMac, Win95 (in development), and of
course Linux. Harald
Deischinger is responsible for the Linux port. He recently released a new
version (0.9) which is available from the following sites: The input to the program can be any audio source, such as a microphone, a
CDROM drive (though you must have the drive connected to your soundcard), or
even a sound file. Cthugha takes the digital audio information and, after
passing the data stream through any combination of filters, displays it to the
screen in real time. The keyboard is used to change the various parameters
either specifically or randomly. The simplest displays resemble the screen of
an oscilloscope being fed audio data (Cthugha has been called "an
oscilloscope on acid") but as more optional filters are added the display
becomes baroquely intricate. If too many filters are active the
resulting images can be chaotic, with little discernible relation to the sound
being processed.
The Linux version of Cthugha is compiled into two executables: cthugha,
which is a console application (using Svgalib), and xcthugha, which runs
either in an X-window or as a full-screen X application using the new DGA
extensions. This last requires XFree86 3.2 or later. Xcthugha can also be
run as a screensaver; in former releases this was a separate executable.
In this release the X11 version runs faster and smoother than in earlier
releases, but I still prefer the console version. It's the quickest and most
responsive of the three interfaces and (in my experience) the only usable
version on a machine less powerful than a mid-range Pentium.
Running Cthugha reminds me of playing a musical instrument. The first
attempts aren't consistently pleasant, but with practice a measure of control
is gained. Orchestral or loud rock music can benefit from
low gain settings, which help to produce a non-chaotic display. The good sort of
recording to start with is music with few voices or tracks. A vocalist with
minimal accompaniment or solo instrumental music give good results
while you gain a feel for the program.
Cthugha comes with several "translation tables"; these are
filters which map the display to various moving patterns, such as spirals or
the appearance of traveling through a starfield. I don't use them much, as it
seems to me they obscure the relationship between the music and the display.
The tables also tend to increase CPU usage. Try them and see what you think,
as they seem to be popular with other Cthugha users.
The other filter categories are more useful. The "wave" filters control the
general shape of the sound waves. These run the gamut from basic oscilloscope
sine and square waves to angular lightning-like patterns or floating clusters
of fire-flies. The "flame" filters add to the waves trailing clouds of glory
(I've always wanted to use that phrase in a non-ironic sense!).
Using a microphone as input is fun, especially if there are kids around.
Seeing your voice represented as billowing clouds of iridescent plasma is
novel, to say the least. Various musical instruments are interesting to try
as well; if one person plays the instrument while the other keys in parameters,
a combination which seems to reflect the character of the melody can often
be found. If you should happen upon a combination of settings which results
in a particularly pleasing screen just press the a key and those
settings are entered into your Cthugha initialization file.
Another option is the Fast Fourier Transform, an algorithm which gives an
entirely different look to the sound; it's hard to describe, but FFT seems
more three-dimensional and subtle. The sampling rate should be reduced to
22000 hz. (from the default of 44000 hz.) since FFT adds one more level of
computation to the sound-translation process.
Kevin Burfitt's decision to use the Fractint 256-color palette file as the
Cthugha palette file format was fortuitous. Over the years Fractint users
have come up with a multitude of palette files among which can be found
palettes to please anyone's taste. The Fractint fractal generator includes a
handy palette-file editor which can be used to create or modify palettes for
Cthugha. I'm not sure if the palette editor is included with Xfractint -- I
mostly use the DOS Fractint in a Dosemu console session.
Here are a couple of screen-shots of xcthugha running in a 320x200
window: These are snapshots, of course, and show little of the dynamic quality of
Cthugha reacting to the music. The above images, by the way,
are of an old recording of Sarah Vaughan singing with piano accompaniment.
Last modified: Sun 27 Apr 1997
After I finish these Gazette articles and get them uploaded to SSC,
I can usually count on a URL changing or a newer version of a program being
released. Sometimes that very day! The Gazette readers are also quick to let
me know of any factual errors I've made. I've accumulated
several of these corrections and updates and shall present them here.
Last month I wrote a short piece about GV, a new Postscript file viewer.
I received a letter from the maintainer of the Debian GV package: FileRunner has been updated several times since I reviewed it several
months ago. The latest version, 2.3, has improved FTP capabilities (including
the option of downloading files with a separate background process). I must
confess I'm addicted to this file-manager. Once you get the hang of it file
manipulation and directory traversals become so speedy that using it as
root can be risky! Check the FileRunner WWW
site for latest releases and news.
Here's an example of a user-configured action-button for FileRunner, which
will mostly interest XEmacs users (though it could probably be adapted easily
for use with GNU Emacs). Create a file in the ~/.fr directory named
cmds, then enter this text into it: For this to work, you must have gnuserv running; this can be started
from your ~/.xemacs-options file by including the line In LG #14 I wrote about the minimalist window-manager wm2, written by
British programmer Chris Cannam. Since then wm2 has spawned a variant, known
as wmx. Evidently Mr. Cannam felt that spartan wm2 was becoming decadently
featureful. Wm2 was stripped down to the bare minimum; no more
frame-background pixmaps,etc. Wmx is just wm2 with the afore-mentioned
pixmaps and a basic virtual-desktop utility. It has one more feature which I
thought was very cleverly designed: if you click the middle mouse button on
the desktop an application menu appears. Unlike most window-managers, the
entries on the menu are a snap to set up. Simply create a subdirectory
stemming from your home directory called .wmx and symlink executables
to it. This can be even done while wmx is running. Whatever appears in
~/.wmx will appear in the menu. The menu can be configured with a
transparent background so that it has a very stylish and spare appearance. As
with wm2 the configuration can only be changed by recompiling, but this can be
done very quickly as the source is not large or complex. Source for either
wm2 or wmx can be obtained from
the wm2 web-site.
A reader pointed out an error in my description of the Afterstep
window-manager in LG #14. Rather than being based on Fvwm2 code, Afterstep is
based on Fvwm version 1 code. Incidentally, pre-release 6 has been released
and is well worth a trial. Several bugs have been fixed but the improved
documentation alone makes it worth the download.
Lately it seems that a fad is sweeping the insular world of vi-like
editors. First the X versions of Elvis and Vim appeared with
pull-down menus; now it appears that Xvile will soon have a menubar as well.
If a: you like vile/xvile and b: you have the Motif libs
installed, you may want to take a look at the patches for vile 7.00 available
from the Vile ftp site.
The patches A through G need to be applied to the vile 7.0 source. It looks
like the menu items will be fairly easy to set up, as they make use of the
standard vile functions. An implementation for non-Motif X setups is
planned.
I have mixed feelings about GUI conveniences such as menus in a vi editor.
One of the appealing traits of these editors is the lack of such visible
features combined with a wide array of invisible and powerful commands.
Little overhead but great power and speed. If you have to reach for the mouse
and select a menu-item, why not use Nedit (for example) which is designed as a
mouse-oriented editor? On the other hand, how many users have had an
unpleasant first-time experience with vi and rejected it forever? At
least the menubar will have a "quit-ZZ" item, allowing a novice to end a
first session without having to desperately flee to another virtual console
and kill the vi process from afar!
The latest version of this versatile desktop/file manager can be found at
the TkDesk home site. Version 1.0b4 has been released and many
minor bugs have been fixed. There are three patches available on the web-site
which should be applied by users of the program. Two of them are changes to
*.tcl files, whereas the third is a c-source-level change which
requires recompilation. Debian users can instead install a patched TkDesk
package which is available from the /bo/binary-i386/x11 directory of
ftp.debian.org and its mirrors.
For the past several months a beta development cycle has been underway in
preparation for the release of mc-3.1.5. The recent releases (the latest as
of this writing is patchlevel 25) have been very stable and usable. If you
use the Midnight Commander frequently it might be worth your while to try the
new version, as many improvements have been made.
An internal editor has been incorporated into mc, though you still can
change the settings and use any console-mode external editor. The FTP
capabilities of mc have been augmented and the Tk version has made great
strides and needs just a few more features to be the equal of the
classic console version. mc now has the ability to dive into *.rpm and
*.deb files in the same manner it has been able to do with *.tgz
and *.zip files, allowing you to inspect their contents without
unpacking the archives.
It's only available in source form, but it comes with a good configure script
and compiles easily here. The source is available from
the mc home site.
Last month I wrote about the release of XEmacs 19.15. The XEmacs team
didn't stop and rest on their laurels (probably because some unexpected
problems showed up after the release!); beta releases of XEmacs 20.1 began
showing up about twice a week at ftp.xemacs.org. It looked as if
version 20.1 was about to be released, but for some reason the release was
cancelled and they moved on to betas of 20.2. I'm running beta 2 now,
and have found that several small problems with 19.15 have been fixed. The
Customization utility works quite a bit better now, for one. When 20.2 is
released I would recommend obtaining it, as it looks like it will be an
improvement over 19.15. Another approach if you've already installed 19.15 is
to visit
the XEmacs patches page, which offers patches to upgrade 19.15 to
patchlevel 2. The problems dealt with are described on the page; if the
patches concern modes or utilities you never use, there's no point in applying them.
Welcome to the Slackware distribution of Linux! This chapter aims
to help the new Linux user or administrator evaluate Slackware,
plan a Slackware system, and install Slackware Linux. In it you'll
find an emphasis on careful planning rather than rushing into
an impetuous installation. A special worksheet is included to
help you "get it right the first time", which I hope will be especially
useful to overworked Unix administrators in busy environments.
Whether or not to choose Slackware as the flavor of Linux you
will use is a serious consideration. It may seem like a trivial
decision now, but Linux boxes have a way of taking on more and
more responsibility in organizational computing environments.
Plenty of Linux experiments have evolved in their first
year to become mission-critical machines serving many more users
and purposes than originally intended. Slackware is one of the
most widely used distributions of Linux. When it comes to finding
the newest, easiest, or most carefully planned distribution of
Linux, Slackware may be "none of the above". Some background on
the life and times of Slackware put things into perspective.
In 1993, SLS created one of the first organized distributions
of Linux. Although it was a great start, the SLS distribution
had many shortcomings (it didn't exactly work, for starters).
Slackware, a godsend from Patrick Volkerding, solved most of these
issues, was mirrored via FTP and pressed onto CD-ROMs the worldwide,
and quickly became the most widely used flavor of Linux. For a
while, Slackware was the only full-featured Linux solution.
Other Linux distribution maintainers, both commercial and nonprofit,
have gradually developed distributions that are also well worth
your consideration.
According to statistics maintained by the Linux Counter Project,
Slackware inhabits about 69% of all machines that run Linux. Slackware
is typically obtained via FTP or CD-ROM and installed on a 486-class
computer running at 66Mhz with about 16 MB of memory and 1050
MB of storage. More information about Linux use and the Linux
Counter Project is available on the World Wide Web.
http://domen.uninett.no/\~hta/linux/counter.html By January 1994, Slackware had achieved such widespread use that
it earned a popular notoriety normally reserved for rock stars
and cult leaders. Gossip spread through the Usenet suggesting
that the entire Slackware project was the work of witches and
devil-worshippers!
"Linux, the free OS....except for your SOUL! MOUHAHAHAHA!"
From: cajho@uno.edu Jokes alluding to RFC 666, demonic daemons, and speculation that
Pat Volkerding was actually L. Ron Hubbard in disguise were rampant
in the threads that followed. The whole amusing incident probably
helped Slackware gain some market share:
I LOVE THIS!!
I was browsing here to figure which version of Linux to install,
but after this, I think that I hve no choice but to install Slackware
now.
From: David Devejian All folklore and kidding aside, Slackware is a wise and powerful
choice for your adventures in Linux, whether you are a hobbyist,
student, hacker, or system administrator in the making.
If you are a system administrator, you may already be dealing
with one or more key servers running Slackware. Unless you have
time to experiment at work, sticking to the tried-and-true distribution
may be the easiest way to go. If you expect to get help from Unix
literate friends and colleagues, better make sure they're running
something compatible-odds are they're running Slackware. Its shortcomings
are widely acknowledged, for the most part discovered, documented
and patched whenever possible. You can put together a Slackware
box, close the known security holes, and install some complementary
tools from the other Linux distributions to create an excellent
Unix server or desktop workstation, all in about half a day.
If you are still undecided whether Slackware is the tastiest flavor
of Linux for you, have a look at the "Buyer's Guide"
published in the Linux Journal, which gives a thorough
comparison and evaluation of each major distribution. For a straightforward
listing of Linux flavors, have a look at the Linux Distribution HOWTO
on the Internet: Nine tenths of wisdom is timing. The right time to set up Slackware
is afteryou've carefully planned the installation and
alternatives in the unfortunate event of a problem. A well-planned
installation of Slackware will repay itself many times over in
the future, when the natural process of Linux evolution leads
you to add disk space, install a newer Slackware release, or jettison
any old, inferior operating systems that may linger on your drives.
Like Unix, Slackware Linux tends to grow like a virus. If you
succeed in getting one Slackware box up and running, you're likely
to start infecting other computers that belong to your friends,
family, and coworkers. When this happens, you'll be grateful that
you at least took the time to think through this first setup-and
so will they!
This section will help you decide...
Linux is a powerful operating system, and with power comes responsibility.
Like Linux, the Slackware release treats you with the respect
you deserve as an intelligent human being. If you elect to wipe
out a few hard drives with a misplaced punctuation mark, so be
it. There are graceful and intelligent front-ends to Linux that
allow the average end-user to get lots of productive work done
without ever delving into the cryptic subtleties of Unix setup
and administration. But there's no such luck for you, the appointed
installation guru. If you're going to install Slackware, be forewarned
that you should know your IRQs from your RS232s and your SCSIs
from your IDEs.
This is an essential element for planning any Linux installation. The only
Slackware-specific hardware issue is this: you must confirm that
the particular version (vintage, release) distribution of Slackware
you'll be installing from provides a kernel and drivers to
support your hardware. You're in great shape with just about
any IBM-compatible personal computer with an Intel CPU older than
the date on your Slackware distribution but younger than 1992
(built after 1992). If you have a bleeding-edge machine, you may
need to download a newer boot disk that includes an updated kernel
and drivers.
For the latest information on it general Linux hardware compatibility,
check the Linux Hardware Compatibility HOWTO document on the World
Wide Web: To check for up-to-the minute Slackware news, such as which boot
kernels are available, you can look in this directory of the Slackware
home ftp site, ftp.cdrom.com: Careful planning of file systems and the storage media upon which
they reside can spare you hours of painful juggling at a later
date. In particular, putting all of your custom administration
files, user homes, and local software onto dedicated partitions
or disks will allow you to upgrade Slackware on the root partition
with minimal disruption to your improvements to the system.
A typical personal computer has one fixed disk drive. If you're
a hobbyist or power user, you may already have installed more
than one Operating System on that drive. For example, your computer
may have shipped running MS-DOS or Windows 95 as a pre-loaded
operating system, after which you added another operating system
such as OS/2, NeXTstep, Geoworks, or Linux. To run multiple operating
systems from one drive, the disk is divided into separate areas
known as partitions. Each partition may contain a different operating
system. Once you've installed a second OS, you also need to install
a small program called a boot manager or OS loader that runs at
system startup time and offers you a choice of all the installed
operating systems.
If you're adding Linux to a computer running a lesser OS, you
may elect to keep the old operating system around for kicks. Take
a look at the Linux Loader (LILO), a high-powered boot manager
that comes free with Slackware. The latest distribution of LILO
and its documentation are available via FTP from this URL: In a simple world, you can set up Linux to run on a single disk
partition (or maybe two-one for swap). In a real-world, multi-user
Unix system, a single-drive file system setup creates unnecessary
risks and hassles you can avoid by distributing the file system
across multiple partitions. It's all the same to Unix, which views
the file system as a continuum of available space comprised of
all the disks and partitions "mounted" into various locations
on the file tree.
If you create a Slackware setup on only one drive partition, you
effectively put all of your eggs in one basket-one user may receive
an abundance of e-mail and overload the /var/mail
file system, another might store enormous files in their home area,
etc. As with many Unix quandaries, you have a choice of solutions
to control file system use, including quotas and user limits. Distributing
your Unix file system across multiple partitions and disks has
an extra benefit for Slackware users-it allows you to upgrade
the Slackware installation with a minimum of pain.
The Linux file system standard puts the personal space of each
user into a subdirectory of /home. The user Linus
would typically have a home under /home/linus, the
user Patricia under /home/patricia, and so on. An
easy way to protect this file system during future upgrades is
to mount /home on a separate disk or partition. Same goes for
custom programs and resources you add to the off-the-shelf version
of Slackware-plan to put these on a separate disk mounted to /usr/local
and you'll have much less grief when it comes time to upgrade.
"Where things go"---or where they try to go unless you dictate otherwise---
in a Slackware box is determined by a standard file system layout,
called the Linux File system Hierarchy Standard. Read all
about it URL: In some settings, Linux boxes are assembled from leftover parts-"worthless"
386 and 486 motherboards, old grayscale monitors, and discarded
hard drives. You may need to link together several ancient 40MB
hard drives to come up with enough space to install Slackware.
In other environments using Linux, there are so many users and
such large development projects that several of the biggest, state-of-the-art
drives or drive arrays must be integrated to provide enough space.
You can install Slackware onto more than one disk at once by
designating individual disks to hold specific parts of the Slackware
installation (just like using multiple partitions), creating a
logically continuous and unified file system.
For an informed second opinion on partitioning, swap space setup,
fragmentation and inode size consult Kristian Koehntopp's Partitions
Mini-HOWTO via Internet URL: 24-Aug-95 NOTE: Trying to upgrade to ELF Slackware from a.out
Slackware will undoubtedly cause you all kinds of
problems. Don't do it.
Patrick Volkerding
One thing we don't hear too often with Slackware is the U-word.
Slackware's setup program is designed to put a fresh operating
system onto empty hard disks or empty disk partitions. Installing
on top of a previous Slackware installation can erase your custom
applications and cause compatibility problems between updated
applications and older files on the same system. When Slackware
was first put together, everyone was a first-time Linux user,
and the system was always experimental-reinstalling the entire
operating system and applications was the norm in a developmental
system. Today, many institutions and businesses now run mission-critical
applications on Slackware Linux. In such environment, a simple
reboot is a planned activity and taking down the system and overwriting
all the user files or custom applications is absolutely unacceptable.
So, if you cracked open these pages to plot an upgrade, better
think twice. If you're planning a first-time Slackware installation,
there are a few decisions you can make now that will ease upgrading
in the future:
Teaching you how to finagle a Slackware upgrade is beyond the
scope of this chapter, but it is workable if you are an experienced
Unix administrator and you've taken the precautions above. There
is an Internet resource that claims to analyze your distribution
and bring it up to date across the Internet, you might want to
have a look at this URL if you're facing an upgrade situation: Or read, weep, and learn from the upgrade expertise of Greg Louis
in his mini HOWTO document: Upgrading Your Linux Distribution,
available where finer LDP publications are mirrored: Slackware can be installed from a variety of media and network
sources to fit your needs and budget. Every installation method
will require you to have at least three floppy diskettes available
to get started.
Installation from CD-ROM is fast, popular, and convenient. Although
someone has to break down and pay for the initial purchase of
a CD-ROM, sharing CD's is encouraged. Because Linux and
the Slackware distribution are copylefted, you may make as many
copies as you like. CD-ROM installation is also a bit better practice
in terms of netiquette, since you're not hogging bandwidth for
an all-day FTP transfer. Finally, you may be grateful for the
extra utilities and documentation that accompany the CD-ROM, especially
if you run into installation hassles or need to add components
in the future.
If you're a hobbyist (or want to watch a few dozen Slackware installs
before taking on the task at work), see if there is a LUG (Linux
User Group) in your area that sponsors install parties. Imagine
a roomful of generous and knowledgeable hackers uniting to share
CD-ROMs and expertise with other enthusiasts.
FTP
According to the Linux Counter Project, FTP is still the most
popular way to obtain Linux by a narrow margin. Once you transfer
Slackware from the closest possible FTP mirror, you'll still need
to put the Slackware 'disk sets' onto installation media such
as a hard drive partition or laboriously copy them onto 50-odd
floppy diskettes.
NFS
In a networked environment, it is possible to install Slackware
on a shared file system and allow everyone on the Local net to
attach to this shared location and install. If you have the technical
know-how or a geeked out system administrator who is Linux-literate,
this is a great way to go. The initial distribution of Slackware
can be added to the network via CD-ROM, FTP, Loading floppies,
tape, or even via a remote NFS share across the Internet! For
details on such a remote share, see these URLs:
Floppy
It's time consuming, but it works-you can buy or create the pile
of floppies needed to install Slackware and then feed them into
your box one-by-one when prompted. Slackware 'disk sets' are actually
designed and arranged to fit floppy diskettes. If you happen to
have a huge stack of recycled high-density floppy diskettes at
your disposal, this can be the most economical way to go.
Hard Disk
This is the way to do it if you've transferred the Slackware distribution
across the Internet via FTP-you'll escape the floppy trap by merely
creating boot, root, and rescue diskettes. It requires you to
have an extra disk or disk partition with extra space to hold
the Slackware files during installation (you can erase them afterwards).
Installation from the hard drive is also a workaround if you bought
the CD but your CD-ROM drive is not supported by any of the Linux
kernels that come with the Slackware CD. You can use your present
operating system to transfer the Slackware files onto spare hard
disk space, then boot into the Slackware installation.
Tape
Still experimental as of this writing, tape offers a great compromise
of speed and economy when installing Slackware-worth considering
if a friend with compatible tape drive can dupe a CD or FTP archive
for you. Get the latest details from the TAPE section of the INSTALL.TXT
file that accompanies your Slackware distribution.
Even if you're gifted with a direct T-3 Internet connection that
allows you to suck up a new distribution of Slackware right off
the 'net, you'll be wise to start by building the two Slackware
setup disks (boot and root) before proceeding. In the event of
an unfortunate accident (power outage, feline friends traversing
the keyboard, or even human error), these two little disks, in
the hands of an experienced Unix hacker, may be able to revive
your system or at least rescue your personal files.
During the installation, must choose which disk sets (Slackware
lingo for collections of software) and individual programs to
install. You can usually just accept the default recommendation
of whether or not a package is worth having. A few setup decisions
are crucial. Mid-installation is no time to decide you
want to boot back into OS/2 and look up what kind of graphics
chip your video card uses, which network card you've got in there,
or whether you'll be needing a SCSI or an IDE kernel to get started.
I've often blurted out to a supervisor, "Oh sure, I can have
it up and running in a few hours." Famous last words.
If anyone else has a stake in the Slackware computer's health,
you owe it to them and yourself to think through a less-than-perfect
installation attempt:
After the files are all copied, Slackware can go on to do most
of the system and network configuration, if you're ready. To help
you plan your decisions, Section 3 consists of a worksheet derived
from the text-based Slackware setup program. You can use this
worksheet to record answers in advance (while your computer is
still working!), so you'll be ready with the necessary details-partitions,
IP addresses, modem and mouse IRQs, host and domain names, and
others that you're required to provide during setup.
Last chance to back out! When using the install from
scratch option,
you must install to a blank partition. If you have not
already formatted it manually, then you must format it
when prompted. Enter I to install from scratch, or
a to add software to your existing system.
Ext2fs defaults to one inode per 4096 bytes of drive
space. If you're going to have many small files on
your drive, you may need more inodes (one is used
for each file entry). You can change the density to
one inode per 2048 bytes, or even per 1024 bytes.
Enter 2048 or 1024, or just hit Enter to accept
the default of 4096. 4096, 2048, or 1024.
These are your Linux partitions (partition list displayed).
These partitions are already in use
(partition list displayed). Enter the
partition you would like to use, or type q to quit
adding new partitions. Use a format such as:
/dev/hda3 or whatever the device name is.
Partition name or quit
DOS and OS/2 Partition Setup: The following DOS FAT
or OS/2 HPFS partitions were found: (partition list displayed).
Good! We're all set on the local end, but now we
need to know where to find the software packages to
install. First, we need the IP address of the
machine where the Slackware sources are stored.
Since you're already running on the network, you
should be able to use the hostname instead of an IP
address if you wish.
There must be a directory on the server with the
Slackware sources for each disk in subdirectories
beneath it. Setup needs to know the name of the
directory on your server that contains the disk
subdirectories. For example, if your A3 disk is
found at /slackware/a3, then you would respond:
/slackware.
These defaults are user definable---you may set any
package to be added or skipped automatically by
editing your choices into a file called TAGFILE that
will be found on the first disk of each series.
There will also be a copy of the original tagfile
called TAGFILE.ORG available in case you want to
restore the default settings. The tagfile contains
all the instructions needed to completely automate
your installation.
You can specify an extension
consisting of a "." followed by any combination of 3
characters other than tgz. For instance, I specify
'.pat', and then whenever any tagfiles called
'tagfile.pat' are found during the installation they
are used instead of the default "tagfile" files. If
the install program does not find tagfiles with the
custom extension, it will use the default tagfiles.
Enter your custom tagfile extension (including the
leading ("."), or just press Enter to continue
without a custom extension.
Tagfile extension Enter
Now put a formatted floppy in your boot drive. This
will be made into your Linux boot disk. Use this to
boot Linux until LILO has been configured to boot
from the hard drive. Any data on the target disk
will be destroyed. Insert the disk and press
Return, or s if you want to skip this step.
Now, we need the domain name. Do not supply a
leading "." Enter the domain name.
Domain name
If you only plan to use TCP/IP through loopback,
then your IP address will be 127.0.0.1 and we can
skip a lot of the following questions. Do you plan
to ONLY use loopback? Yes or No.
Enter your IP address for the local machine.
Example: 111.112.113.114. Enter the IP address for this
machine (aaa.bbb.ccc.ddd).
IP address
You may now reboot your computer by pressing Ctrl+Alt+Delete.
If you installed LILO, remove the boot disk from your computer
before rebooting. Don't forget to create you {/etc/fsta
if you don't have one!
If you've taken the time to plot and plan as recommended in the
preceding sections, then the actual installation will be a piece
of cake. There isn't much writing needed to explain the actual
process of loading Slackware onto your computer(s). You just follow
the steps to build boot and root diskettes, then answer a long
series of questions asked by the menu-driven Slackware installation
program. If you've completed the Slackware Installation Worksheet,
these questions will be familiar and everything will run smoothly.
Choose Your Kernel
When installing Slackware Linux, you must create a boot diskette
with a Linux kernel that is specially prepared to recognize your
system hardware. For example, to install Slackware from an IDE
CD-ROM drive onto a SCSI hard drive, the kernel that you put onto
the boot diskette will need to have drivers for your SCSI card
and your IDE CD-ROM drive.
The kernels are stored as compressed binary image files
that you can access from most any operating system to create a
Slackware Boot diskette. On the Slackware FTP site, CD-ROM, or
NFS mount, you'll find a subdirectory called bootdsks.144-containing
1.44 MB kernel images for creating boot disks on 1.44MB high density
3.5'' floppy diskettes. If you're working from a 5.25'' floppy
diskette drive, look in a directory called bootdsks.12 for
kernel images that will fit the smaller diskette format.
Table 2 provides a quick reference of the kernel images available
as we went to press. Information and up-to-date boot disk image
information is available from this URL:
ftp://ftp.cdrom.com/pub/linux/slackware/bootdsks.144/README.TXT Table 1
Table 2
Table 3
Unix Operating Systems
If you have the Slackware kernel images on a Unix host that has
a floppy drive, you can quickly create the necessary boot and
root diskettes using Unix commands. You can use the dd
command. The example below which puts the scsi.s
boot kernel image onto the floppy device rfd0:
dd if=scsi.s of=/dev/rfd0 obs=18k
You'll need to repeat this process with one of the root disk images
onto a second floppy diskette.
DOS, OS/2, MS-Windows 95 \& NT
Slackware bundles a utility called rawrite.exe that
will generate boot and root diskettes under DOS-literate operating
systems. To write the scsi.s kernel image onto the
formatted, high-density diskette in your A:$\backslash$ diskette
drive, issue the following command:
RAWRITE SCSI.S A:
You'll need to repeat this process with one of the root disk images
onto a second floppy diskette.
Here's the big anticlimax. After all this planning, preparation,
and partitioning, you're in the home stretch. Make sure the boot
floppy is in the diskette drive, and restart your computer. Now
is a good time to go get some coffee (or whatever you like to
keep you company) and return to the machine ready to play the
part of a button-pushing drone, answering yes-no questions for
an hour or so.
Log in as root (no password) and type setup
or setup.tty
Slackware comes with two versions of an excellent setup program.
One is a colorful, dialog-based, menu-driven version. An alternative
setup, setup.tty, is a text-only version of the installation
that you may actually prefer, because detailed diagnostics and
error messages will stay on the screen and not be erased by the
next dialog box, which happens in the color version. If you're
attempting a Slackware setup on sketchy hardware, I strongly recommend
the less colorful setup.tty routine. If you don't
know much about Unix and would feel more comfortable with an attractive.
``clean'' interface to the same setup process, then by all means
go for the beautiful setup.
Hint: If you have trouble using the arrow keys on your keyboard,
you can use '+', '-', and TAB instead. Which option would you like?
To transfer Slackware onto your system from here should involve
little more than selecting what you want off the menus. By filling
out the Section 3 worksheet in advance, you should be able progress
quickly through each menu in order, until you reach the INSTALL
option, at which point things may s l o w down: you are advised
to select the PROMPT feature and read about each software
package, deciding whether or not you'd like it to end up on your
Slackware system. The last part of a regular setup is the CONFIGURE
section on the setup menu, and the questions you must answer bear
a striking resemblance to the second half of the Section 3 worksheet.
Definitely not! At this point, you've either got some annoying
obstacle that is preventing the setup from completing, or more
likely, you're looking at the root prompt
darkstar\~\# Well, if you're plagued by problems, you'll want to proceed directly
to the next section on troubleshooting. If things appear to be
in working order, you've still got some details to attend to.
Sort of like purchasing a new automobile-after you've selected
an paid for a new car, there are still some things you need before
you can drive it with confidence-insurance, a steering wheel club,
and perhaps some luxuries that make the driving experience closer
to Fahrvergn\ügen than FAQ!
Not every Slackware installation is born on cue to expecting system
administrators. I've pulled a few all nighters, sitting down after
work one evening to upgrade a Slackware box and still there struggling
to get the damn thing back online at dawn, before people start
bitching about their missing mail and news. This section will
look at a few common Slackware setup problems, solutions, and
where to look for additional assistance.
Patrick Volkerding, the father of Slackware, has dealt with the
many questions of new users by listening, answering, and anticipating
repeat queries. To catch the new Slackware users before they ask
the same question for the 5,000th time, Patrick has kindly created
documentation and included it with the Slackware distribution.
Three files that you may find very helpful in answering your initial
questions are FAQ.TXT, INSTALL.TXT, and BOOTING.TXT.
Web Support For Slackware
At this time, the Slackware-specific help you'll find on the Internet
tends to be highly customized---such as how to NFS-mount the distribution
on computers within a certain university or how to wire your dorm
room into a particular residential WAN using Slackware.
Usenet Groups For Slackware
The comp.os.linux.* hierarchy of the Usenet is a
treasure-trove of Linux information, not necessarily Slackware-specific.
At present, 11 separate Linux forums handle a high volume of discussion
in this hierarchy. Dozens of other general-Unix newsgroups are
also available. Some discussions relevant to getting Slackware
up and running are:
A group established for figuring out Linux installation and system
administration. The best place to look for clever setup strategies
and to network with others who may have recently installed Slackware.
A must-read for Linux administrators and enthusiasts, C.o.l.a
is a sort of daily Linux digest for the Internet community. The
group is moderated, so only the relevant material makes it into
circulation. The newsgroup is designed as a low-traffic alert
service for announcing Linux-specific software, documentation,
and security warnings.
Here's where to find (or post) the latest Linux FAQs, How-Tos,
READMEs and other documents that answer questions about Linux.
Mail Lists For Slackware
At this time, there are no electronic mail discussions devoted
to Slackware per-se. You can participate in some excellent Linux-related
talk via e-mail, try www.linux.org and asking in the newsgroups
for a few good subscription lists.
Commercial support for Linux is available from some of the CD-ROM
vendors and a long list of Linux Consultants, who can be contacted
through the Linux Commercial and Consultants HOWTO documents:
http://sunsite.unc.edu/LDP/HOWTO/Consultatns-HOWTO.html Don't rest on your laurels quite yet. Especially if your Slackware
machine is a shared computer or lives in a networked environment.
Grooming a computer for community and network use is a bit more
demanding than just running the setup program and
forgetting about it. We'll leave you with a few pointers to securing
and sharing your new Slackware system.
I know you just sat through what may have been a long and perplexing
installation session. But before you move into the house you just
built, consider tearing it down and starting over again. Friedrich
Nietzsche had a quote:
"A man learns what he needs to know about building his house only
after he's finished."
If, in the process of installing the system, you had some thoughts
about how you might do it differently, now is the time. If your
Slackware Linux box will be a multi user machine or a network
server, there may never be such a convenient opportunity to reinstall
or reconfigure the system in radical ways.
Before you put away the CDROM or return the 50 floppy disks you
borrowed to run the Slackware installation, sit down and test
each application that your users may expect to find in working
order. If professor Bien absolutely has to have emacs humming
under X-Windows, you'd better test it out now, while you've still
got the workstation 'in the shop.'
Did you set up this Linux box to serve a specific purpose in your
organization, such as...
Get Off The LAN At Once
Out of the box, Slackware is an insecure system. Although Patrick
does his best to create a secure distribution, a few inevitable
holes become known, and patches or workarounds are made available
in the system administration (and cracker) communities. If you
installed Slackware from a network source such as an NFS-mounted
drive, you should temporarily disconnect your box from the LAN
after a successful installation, while you plug a few holes.
Give Root a Password
By default, a new Slackware box will not require a password for
the root user. When you're comfortable that your
new Slackware system is stable (after a few hours, not days or
weeks), add a password to protect the root account.
Login as root and type:
passwd root
Give Yourself An Account
On large shared systems, the super-user root account is not used as a
working login account by any individual. If you're interested in
system administration or are running a networked machine, this is a
good precedent to follow. Use the \texttt{/sbin/adduser} program to
make yourself a login account, rather than working out of the root
login. I always smile when I see students and hobbyists posting
proudly to the Usenet as root@mymachine.mydomain. Be humble and safe,
create another login account for your daily work and use su (rather
than login) to enter the root account sparingly.
Deny Root Logins
Not only is it uncommon to work as the root user,
it is not considered secure to login as root across the network.
Administrative users usually connect to a Unix box as their regular
username login, and then use the su utility to become
the root user as needed. To prevent crackers, hackers, and ignorant
users from logging in directly as root, edit the file /etc/securetty
and comment out (prepend a pound \# sign before) all but the local
terminals:
console
tty1
tty2
\# ttyS0
\# ttyS1
After this fix, users who attempt to login in as root across
the network will be denied:
Linux 2.0.29 (durak.interactivate.com) Apply the Simple Fixes
Slackware installs itself with some very real security problems.
Rather than master Unix security and sleuth out these vulnerabilities
yourself, you can jump start the hole-patching process by visiting
a web resource maintained for just this purpose, called Slackware
SimpleFixes:
http://cesdis.gsfc.nasa.gov/linux-web/simplefixes/simplefixes.html Check For Patches On ftp.cdrom.com
As an actively maintained Linux distribution Slackware updates
and patches are available from:
ftp://ftp.cdrom.com/pub/linux/slackware/patches/ Stay Current
You might like to subscribe to one or more electronic mail lists
that alert users to issues in Linux administration, such as:
Like how things are running? Save it for a rainy day by backing
up. Amanda (The Advanced Maryland Automatic Network Disk Archiver)
is one of several backup options for Linux installations. You
can learn more about Amanda from:
http://www.cs.umd.edu/projects/amanda/index.html/
It all started with a simple question: `Why don't we install
Linux at all our meetings instead of at only some of them?'
The North Texas Linux Users
Group had been in existence for only about five months,
and we wanted to make sure to spread the word in North Texas
about Linux. We wanted to educate the computing public in our
area about Linux, but we also wanted to let experienced
computer users know that Linux could handle anything they
threw at it.
After meeting at Texas Christian University for our first few
meetings, we signed a contract with the DFWXchange that enabled NTLUG
to meet at the Dallas
Infomart. The DFWXchange is an umbrella organization that
allows Dallas-Fort Worth users groups to meet at the Infomart
for free, with all costs being absorbed by the many commercial
vendors who also meet at the Infomart during the Super Satuday
Sale. So every month between 3,000 and 5,000 computer users
from the Dallas-Fort Worth Metroplex converge on the
Infomart---the premier meeting facility in the Southwest
devoted exclusively to computer and technology events and
organziations---to participate in users groups meetings and to
take advantage of some really good prices on computer-related
hardware. It's a big party.
It didn't take long for the NTLUG leadership to realize that
we had stumbled upon a great opportunity: we wanted to let
computer users in our area know about Linux, and we were now
meeting every month in a facility filled with thousands of
potential Linux converts. Our solution was to start the
Linux Installation Project, which we call the
LIP.
The goal of LIP is simply to install Linux on as many
computers as possible. Those of us who participate in this
project month-to-month have discovered that the very best way
to advocate the use of Linux is to make it easy and painless
for the unconverted to do just that: namely, run Linux on
their computer of choice. In other words, Linux is its own
best advocate. After a few weeks without a crash, most people
say goodbye to Windows 95 with zealous enthusiasm. We like to
think of LIP as an ongoing Linux Installation Festival
that allows us to convert computer users to Linux one at a
time.
The first step in establishing LIP as a well-run,
consistent endeavor was to find someone to lead the
effort. NTLUG is fortunate to have a technologically advanced
membership, and it was fairly easy to find someone to lead the
LIP; in fact, we found two such people: Mike Dunn and
Bill Petersen, both of whom are experienced Unix and Linux
Sysadmins. Under their guidance, and due to the generosity of
NTLUG members, NTLUG's LIP has solicited and organized
enough computer hardware to perform up to many simultaneous
installations of Linux by all the usual methods, although
we've found that cdrom installations are usually the most
trouble free.
The word has now spread in and around the Dallas-Fort Worth
Metroplex---from schools and universities to computer vendors
and other users groups---that NTLUG's LIP is the place
to go for a painless installation of Linux onto PCs, laptops,
servers, and even Alpha platforms. We have expanded our
efforts at the LIP booth to include Linux advocacy,
advertisement for Linux vendors who supply us with materials,
the sale of Linux CDs (thanks to Bradley Glonka at Linux Systems Labs), and even
basic Linux system administration and maintenance. We also
spend a lot of time explaining to the uninitiated masses what
makes Linux free and what makes it so much fun.
While we have been happy with the results so far, the
LIP has more work to do. We want to expand our sales
efforts to include other kinds of Linux merchandise (the sales
of which go to support NTLUG and LIP), and we'd also
like to expand our hardware assets to enable more simultaneous
installations. Finally, we also want to develop our users
group assets to such an extent that we can go to other
DFW-area computer events and setup Linux installation and
advocacy booths. NTLUG's approach to the Linux Installation
Project can be summed up in the phrase: "Linux is free. Life
is good."
If you want to learn more about the North Texas Linux Users
Group or our Linux Installation Project, or if you're a Linux
Users Group and would like to talk about setting up your own
local version of LIP, please visit the NTLUG website or contact me
at kclark@computek.net.
Finally, I would be guilty of ingratitude if I did not thank
the following people and organizations that have made the
LIP possible. Please forgive me if I've forgotten
anyone. It's just about impossible not to meet great people
when you work with Linux.
Thanks to all our authors, not just the ones above, but also those who wrote
giving us their tips and tricks and making suggestions. Thanks also to our
new mirror sites. And many, many thanks to Amy for doing most of the work
this month.
Seattle--always a wonderful event. Riley and I spent the morning working
in the yard, clearing out a flower bed that was overgrown with grass. It
felt like hard physical labor after sitting at a desk all week. We rewarded
ourselves by taking a spin on the motorcycle along the Sound. Even as
passenger there is something about riding on a motorcycle that puts a smile
on my face. I guess it helps that I have complete trust in Riley's driving
abilities.
Afterward I went to the Opera House for a talk about Il Trovatore by
Giuseppe Verdi (or Joe Green, as Riley likes to call him). I had seen
the opera itself on Wednesday night--a silly
story as usual but, oh, such
wonderful music! I think it has to be one of my favorites. At any rate
the talk was informative and fun and made a nice end to a very wonderful
day.
Have fun!
Marjorie L. Richardson
Linux Gazette Issue 17, May 1997, http://www.ssc.com/lg/
first parameter - number of bytes
second - begin of output names, e.g. p1
third - name of file to be chopped
#!/bin/bash
echo "* Begin of procedure Chop *"
date
# rm alte Datei
if test -e /usr/TFH/EXAMPLE
then rm /usr/TFH/EXAMPLE
fi
# Test auf Parameter
if test $# -lt 3
then echo "Incorrect number of parameters !
Please repeat procedure call !"
echo "* End of procedure Chop (error) *"
exit 1
else echo "Call was ok"
fi
#
BY=$1
ANZZ=$[(($BY / 20480) + 1)]
quantity=$ANZZ
i=1
recs=0
while test "$i" -lt "$quantity"
do
echo dd if=$3 of=$2_$i bs=1024 skip=$recs count=20 >> /usr/TFH/EXAMPLE
echo sz $2_$i >> /usr/TFH/EXAMPLE
echo rm $2_$i >> /usr/TFH/EXAMPLE
i="`expr $i + 1`"
recs="`expr $recs + 20`"
done
echo dd if=$3 of=$2_$i bs=1024 skip=$recs >> /usr/TFH/EXAMPLE
echo sz $2_$i >> /usr/TFH/EXAMPLE
echo rm $2_$i >> /usr/TFH/EXAMPLE
#
echo "* End of procedure Chop (ok) *"
#
Using ftp Commands in Shellscript
Using FTP as a shell-command with ftplib
ftpget sunsite.unc.edu -r /pub/Linux ls-lR.gz
This command reads the file /pub/Linux/ls-lR.gz from sunsite.unc.edu
Likewise there are other commands with the lib: ftpdir ,ftpsend, ftprm
Everybody tired of typing ftp... every evening to
get the latest patches or whatever. Everyone who is regularly copying
with ftp the same Datafiles.
Of course you can add it to you own application but more experienced users don't
have to use these r-commands anymore. An ftpd is available for the
majority of systems so it is easier to access more of them.
Of course, for any ftp session you need a user/paswdr. I copy into
public area using anonymous/email@ others will need to surly a
password at login, what is not very useful for regular jobs or
you have to use some kind of public login but still I think it's
easier and better to use than the r-cmds.
ACSII-Artwork Translator
I think you'll find the results to be pretty amusing, and slightly
more interesting than the usual bag of HTML table-tricks.
%{
/* Ascii-to-Table version 2.0
**
** A conversion utility to convert gifscii type ASCII-Artwork into
** grayscale HTML 3.0 compliant html documents using tables.
**
** Copyright(C) 1997 by Patrick J.M. Keane -- All rights reserved.
** (pkeane@wilkes.edu)
**
** This program is free software; you can redistribute it and/or modify
** it under the terms of the GNU General Public License as published by
** the Free Software Foundation; either version 2 of the License, or
** (at your option) any later version.
**
** This program is distributed in the hope that it will be useful,
** but WITHOUT ANY WARRANTY; without even the implied warranty of
** MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
** GNU General Public License for more details.
**
** You should have received a copy of the GNU General Public License
** along with this program; if not, write to the Free Software
** Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
**
*/
#include
",
((shade1set==0) ? value : shade1),
((shade2set==0) ? value : shade2),
((shade3set==0) ? value : shade3)) ;
printf(" ") ;
}
main(int argc, char *argv[]) {
int c;
extern int optind;
extern char *optarg;
extern int opterr;
while ((c = getopt(argc, argv, "w:r:g:b:xh")) != EOF) {
switch (c) {
case 'x':
reverse = 1 ;
break;
case 'h':
fprintf(stderr, "Usage:\n\tcat asciifile | a2t [-h] [-x] [-[rgb] value] [-w width] > document.html\n\n") ;
fprintf(stderr, "\t-h : This help screen\n") ;
fprintf(stderr, "\t-x : Reverse output\n") ;
fprintf(stderr, "\t-r value : Constant R ", width) ;
else
printf(">") ;
printf ("
\n") ;
yylex() ;
printf(" ") ;
printf(" \n") ; }
" " { maketd("00") ; }
. { fprintf(stderr, "Warning: Character %s is not recognized.\n",
yytext) ;
fprintf(stderr, "Choosing a medium color!\n") ;
maketd("97") ; }
%%
void yyerror(char *msg) {
fprintf(stderr, "^GError :\tLine %d: %s at '%s'\n", yylineno, msg, yytext) ;
}
int yywrap() {
return (1);
}
Including Graphics in Linuxdoc SGML
Date: Thu, Apr 17, 1997 at 07:48:19PM +0200
You can already include PostScript images in Linuxdoc-SGML which
will get included in TeX output (and consequently in DVI and
PostScript). Linuxdoc-SGML doesn't support images for HTML,
however.
X Configuration Issues
From: Michael J. Hammel, mjhammel@emass.com
Multiple X Displays
From: Michael J. Hammel mjhammel@emass.com
Color Depths with X
From: Michael J. Hammel mjhammel@emass.com
After fiddling with the xf86config file in a concerted effort to coax X
into displaying 16 bit color, I was dismayed to learn that with my
current hardware (16 megs RAM and a Cirrus Logic GL-5426) 16 bit color
is *impossible*...not because of any hardware in-capability, but because
of a certain limitation of X Windows itself...a problem with linear
addressing. Seems that to have 16 bit color under X, one must have
linear addressing enabled, which only works if the system has *no more
than 14 megs RAM*.
Figuring Out the Boot Process
From: David Ishee dmi1@ra.MsState.Edu
One of the things that is confusing about Linux at first is which files
Linux uses to load programs and get the system started at bootup. Once
you figure out which programs are run during the boot process, which
order are they run? Here is an easy solution.
edit /etc/rc.d/rc.sysinit and add the following lines at the beginning
echo " "
echo "**** Running /etc/rc.d/rc.sysinit ****
echo " "
ftping Home
From: Kaj J. Niemi, kajtzu@4u.net
I read your article about ftping home with dynamic IPs.. Here's
something you might need if you get tired of looking at the screen every
time you want to find out the IP.
ADDRESS=`/sbin/ifconfig | awk 'BEGIN { pppok = 0}
/ppp.*/ { pppok = 1; next }
{if (pppok == 1 ) {pppok = 0; print} }'\
| awk -F: '{print $2 }'| awk '{print $1 }'`
This page maintained by the Assistant Editor of Linux Gazette,
gazette@ssc.com
Copyright © 1997 Specialized Systems Consultants, Inc. "Linux Gazette...making Linux just a little more fun!"
Contents:
News in General
GLUE Announcement
Every GLUE User Group To Receive Free Copy of BRU 2000 Backup And Restore
Utility
SOLID desktop for Linux offered free of charge to developers
Solid Information Technology Ltd, http://www.solidtech.com.
The Elsop Webmaster Resource Center
http:www.elsop.com/wrc/
Linux Jokes Wanted
Too Good Not To Print
New User's Group in Knoxville
AfterStep Themes Page
or
http://www.mindspring.com/~xwindow
Version 7 of Corel's WordPerfect for Linux
Computer Comparison
This web site is maintained by Karl Unterkofler, and has comparisons of
various computers running the latest versions of Mathemetica. Karl and
others run a series of tests on the machines, that involve timing
mathematical problems.
Word Processor for the Linux Environment
Software Announcements
Xcoral 3.0
Beta Version of EM86
XForms V0.86
Debian 1.3 Available for Beta Test
Freedom Desktop Lite Announced (1.01)
For more information and the ftp site feel free to visit http://freedom.lm.com/desktop.html
This page written and maintained by the Editor of Linux Gazette,
gazette@ssc.com
Copyright © 1997 Specialized Systems Consultants, Inc. "Linux Gazette...making Linux just a little more fun!"
The Answer Guy
By James T. Dennis
jimd@starshine.org
Starshine Technical Services,
http://www.starshine.org/ Contents:
fs's
"What were the disaster plans, and why
are those plans inadequate for this
situation?"
dd if=/dev/sda | od | less
dd if=/dev/sda of=/dev/st0
Linux/Unix Emulator
Using X with 2 Monitors and 2 Video Cards
I was wondering how I would go about using X with 2 monitors and 2
video cards? I am currently using XFree86 window manager. I know you
can do this with the MetroX window manager but that costs money :(.
Virtual Hosting
tcpd has supported virtual hosting for more than two years. Below
is a fragment from the hosts_access(5) manual page.
In order to distinguish clients by the network address
that they connect to, use patterns of the form:
process_name@host_pattern : client_list ...
Response from Weitse Venema
Do you know where there are any working examples of this
and the twist option posted to the 'net? I fight with
some of these and don't seem to get the right results.
Automatic File Transfer
In Linux Gazette, there is a mention of how to transfer files
automatically using ftp.
Here is how:
#!/bin/csh
ftp -n remote.site << !
user joe blow
binary
put newfile
quit
!
Installing wu-ftpd on a Linux Box
I just installed wu-ftpd on my linux box. I have version 2.4.
I can login under one of my accounts on the system and everything
works just fine.
#! /bin/sh
exec strace -o /tmp/ftpd.strace /usr/sbin/wu.ftpd
Trying to Boot a Laptop
I've got a Toshiba satellite pro 415cs notebook computer on which I've
installed RedHat 4.1. RedHat 4.1 was installed on a jaz disk connected
via an Adaptec slimscsi pcmcia adapter. the installation went
successfully, i believe, up until the lilo boot disk creation. i
specified that i wanted lilo on a floppy - so that nothing would be
written to the internal ide drive and also so that i could take the
installation and run it at another such laptop. after rebooting, i
tried booting from the lilo floppy that was created, but i get nothing
but continuous streams of 0 1 0 1 0 1...
@ECHO OFF
ECHO "About to load Linux -- this is a one-way trip!"
PAUSE
LOADLIN lnx2029.krn root=/dev/sda1 ro
LOADLIN lnx2029.krn single root=/dev/sda ro
yash
zmodem Reply
From: Donald Harter Jr., harter@mufn.org
I saw your post about zmodem in the Linux Gazette. I can't answer the
readers question, but maybe this will help. My access to the internet is a
dial in account(no slip, no ppp). I access the freenets. I can't use
zmodem to transfer files from the internet and freeenets to my pc. I can
use kermit though. It seems that there are some control characters involved
in zmodem that prevent it from being used with my type of connection. I saw
a some information about this on one of the freenets. They suggested using
telix and another related protocol. I tried that, but it didn't work
either. Kermit is set up to run slow. You can get kermit to go faster in
certain circumstances by executing its "FAST" macro. I can download data at
about 700cps with the "FAST" macro of kermit. Unfortunately kermit hangs up
the line for me so I have to "kill -9 kermitpid" to exit it. That problem
can probably be eliminated with the right compile options. In certain cases
I can't use the "FAST" macro when uploading.
StartX
Hi, I was wondering if you can help me out. When I use the command
'startx -- -bpp16' to change the color depth, the windows in X are much
bigger than the monitor display. So, nothing fits properly and
everything has become larger. But the color depth has changed
correctly. I use FVWM as my display manager. Is there some way to fix
this problem?
IMAP and Linux
Being a big IMAP fan (and glad to see it finally getting recognition:
Netscrape 4 and IE4 will both support it), your answer left a lot out.
More IMAP
PINE - one of the easiest to use mail clients around - does IMAP just
fine. You can read mail from multiple servers and mailboxes and save
it locally or in remote folders on the servers - which is what IMAP is
all about: Internet Message Access Protocol = flexible and
configurable *access* to mail servers without having to pop and fetch
messages all over the place (but still having the ability save locally
if you want).
strace -o /tmp/strace.script /usr/bin/script
stracetel stream tcp nowait root /usr/sbin/tcpd \
/usr/bin/strace -o /root/tmp/t.strace /usr/sbin/in.telnetd
telnet localhost stracetel
UUCP Questions
I had a couple minor questions on UUCP. If you have a few minutes, I'd
appreciate the help immensely. I'll tell you a little bit about what we're
doing.
Unfortunately I this question is too broad to answer
via e-mail. O'Reilly has a whole book on uucp and
there are several HOW-TO's for Taylor UUCP and
sendmail under Linux.
jim%mercury@starshine.org
Using MS-DOS Floppies
Hi, I have a problem that I can't find the solution to:
# /etc/fstab
/dev/fd0 /mnt/tmp umsdos noauto,rw,user 0 0
#! /bin/sh
/bin/mount /mnt/tmp
/bin/sync
/bin/umount /mnt/tmp
-r-sr-x--- 1 root disk .... /bin/mount
inetd Questions
Answer guy,
I have two questions for you.
/usr/sbin/inetd /etc/inetd.fred 192.168.14.0 17.18.0.0
/usr/sbin/inetd /etc/inetd.barney barneyweb
/usr/sbin/inetd /etc/inetd.wilma 192.168.2.3
inetd.conf.207.122.3.8
inetd.conf.207.122.3.90
Navas Modem FAQ
The next time you answer a modem question, you'd do well
to recommend reading of the very good Navas Modem FAQ at
http://www.aimnet.com/~jnavas/modem/faq.html/
Setting Up a Modem
linux +aztech
User Identification
i need your help. for some reasons i have to identify a user on my
webserver by his/her ip-address. fact is that users logon comes from
different physical machines. that means that i have to assign something
like a virtual ip-address to a users log name. something like a reversal
masquerading.
Duplicating a Linux Installed HD
I just read your response to duplicating a hard drive using dd.
I think using dd limits the uses of this technique too much.
tar -c -X /tmp/excludes -f / | (cd /mnt; tar xvf -)
The file....
/tmp/excludes should contain:
/mnt
/proc
and any other non-local, mounted drives, such as nfs mount points.
find ... -xdev -type f | tar cTf - - | \
(cd ... && tar xpf - )
Another is to use:
find ... | cpio pvum /new/directory
... which I only learned after years of using
the tar | (cd ... && tar) construct.
tar tf /dev/st0 | gzip > /root/tapes.contents/.....
antares-X.19970408
find / -newer /root/tape.contents/.... \
| egrep -v "^(/tmp|/proc|/var/spool/news)" \
| tar czTf - /mnt/mo/diff.`date +%Y%m%d`.tar
Copyright © 1997, James T. Dennis
Published in Issue 17 of the Linux Gazette May 1997
"Linux Gazette...making Linux just a little more fun!"
CLUELESS at the Prompt: A Column for New Users<\H2>
By Mike List,
troll@net-link.net
Welcome to installment 4 of Clueless at the Prompt:
a new column for new users.
Connecting to a Second ISP...or Third, or
I recently got e-mail from a guy who wanted to know how to connect to a
second ISP.His e-mail address apparently wasn't valid, and it got bounced
back several times. Just as well, since I didn't have the answer at that
point. Well, I got this idea, and I tried it and it works.Here's the deal:
First,
cp /usr/sbin/ppp-on to /usr/sbin/ppp-on.anysuffix
Dealing With a Dynamic IP
ifconfig
nslookup the.num.ber.
Name: your Fully Qualified Domain Name
Address: IPa.ddr.ess.!!!
hostname Name
Using your Domain Name
If you have a shell account at a computer located at a university or
school near you, this will amaze you. Oh yeah not, by way of a plug, but
there is a semi commercial telnet box called linuxware.com(you will have
to look up the URL yourself semi- plug, you might say, I'm a subscriber)
What am I talking about? Using X to run apps from the remote computer on
your screen. You can actually run a program that isn't installed on your
computer, in X with the remote computer supplying the program. I think
it embodies the essence of networking, with permissions set right, you
can co-author a document, play a multi-user game (MUD)use a
talkprogram, like ytalk, or do office or school work from your home
computer.Here's what you need to do. First, you need to know and have
your FQDN listed by typing:
hostname
xhost + the.telnet.box
telnet FQDN or ftp IPa.ddr.ess.!!!
Just a Reminder: Read the whole Linux Gazette
Formatting Floppy Disks in Linux
Other Stuff I've Collected/Found out Since Last Time
find
locate /filename |less
whereis filename
Next Time- Let me know what you would like to see in here and I'll
try to
oblige just e-mail troll@net-link.net
me and ask, otherwise I'll just write about what gave me trouble and
how I got past it.
Copyright © 1997, Mike List
Published in Issue 17 of the Linux Gazette, May 1997
By Jesper Pedersen,
blackie@imada.ou.dk
Intro
What should be done if the user asks to go to the next line at the of
a file?
Since the program works without a GUI, the standard method for
configuring such options is to use a dot-file. In this file, you
may program, which method you will use.
There are two logical
possibilities:
Instead of implementing only one of the solution the people behind Emacs,
have chosen to implement both, and let you decide which one you prefer.
The configurations above may easily be done with a GUI, with the following
widgets in order: A check box, an entry and a pull down menu.
This is exactly what is done in TDG.
which editor would you like to use: emacs, jed, vi or vim
The basic concept of TDG
In region 1, the actual configuration is located. Region 2 is the help
region. In this region help for the whole page is shown, when the window is
displayed. It's also here, help for the individual configuration is
shown, when you press the right mouse button on one of the widgets.
In the Setup->Options menu, you may select which of the
three methods above will be used.The configuration widgets
The ExtEntry widget
Figure 3
The FillOut widget
The Command widgets
Save, Export and Reload
Here's another difference between the save-files and the export-files: You
cannot merge with save-files. This means that if you have a save-file,
which you wish to merge with, you first have to load it, export it, and
then you can merge with it.
The End
If you have some spare time, I would very much like to encourage you to
develop a module for your favorite program. On the home page of TDG, there
is a link to a document, which describe how to create a module for TDG. Send me a letter, and I will be happy
to help you get started with it.
Jesper Kjær
Pedersen <blackie@imada.ou.dk>
Last modified: Wed Feb 5 15:59:35 1997
Copyright © 1997, Jesper Pedersen
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
Set your browser to the width of the line below for best viewing.
© 1997 by
mjh
muse:
elcome
to the Graphics Muse! Why a "muse"?
Well, except for the sisters aspect, the above definitions are
pretty much the way I'd describe my own interest in computer graphics:
it keeps me deep in thought and it is a daily source of inspiration.
his
column is dedicated to the use, creation, distribution, and dissussion of
computer graphics tools for Linux systems.
What a month. Actually, two months. Last month I was busily working
on putting together an entry for the IRTC using BMRT. At the same time
I was trying to teach myself enough about the RenderMan Interface
Specification to put together the second of three articles on BMRT.
I didn't succeed in the latter and ended up postponing the article
by one month. Because I did this
I was able to focus more on learning the interface and worry less
about writing. I think this strategy worked. The scene I rendered for this
months IRTC is the best I've ever done and I managed to gain enough
experience to write a meaningful article on the RenderMan Shading Langauge.
One of the reasons I enjoy doing this column is because it exposes me to
all sorts of people and software. The world of computer graphics for
Linux is constantly growing and the software constantly improves. I
hear about new products just about once or twice a week now and I hear
about updates to existing packages all the time. Its very difficult to
keep track of all the changes (and the fact that I haven't made any updates
to the Linux Graphics mini-Howto in some time reflects this) but I enjoy
the work.
Since things change so often I have found its never clear how many
announcements I'll have for any one month. Its gone from famine to
feast - with this month being the feast. Most of the announcements in
this months column are from April alone. I don't know what happened -
maybe all the bad weather around the globe kept people inside and busily
working and now that the suns out they're starting to let loose what
they've done. I only wish I had the time to examine everything, to play
with them all. But my employer would rather I finish my current project
first. Has something to do with keeping my salary, so they say.
In this months column I'll only be covering two related items. The
first is a case study on learning to use BMRT. When you submit an image in
the IRTC you are required to submit an ASCII text file describing your
image and, to some extent, how you created it. Some people don't put much
work into this. I just about wrote a book. Since the information I
provided covered more than just BMRT I thought it would be relavent to this
column.
The second item is the long awaited (well, I waited a long time to
finish it anyway) 2nd article on BMRT that covers the RenderMan Shading
Language. I think this article came out quite good. I've included quite a
few samples and some general explanations on what they do. I want to
say right up front that I couldn't have done this without lots of help from
BMRT's author, Larry Gritz at Pixar. He was a very willing teacher and
critic who offered many tips and ideas for my IRTC entry. Most of that
also ended up in this article. Many thanks, Larry.
I know I said I'd do an HF-Lab article this month too, but that IRTC
entry took more time than I expected. It was quite addicting, trying
to get things just right. I have started to review HF-Lab once again
and will make it my first priority for next months column. I've already
figured out how to use the output from HF-Lab to produce height fields with
BMRT. Its quite simple really. Anyway, I hope you enjoy this months
articles.
Note: I've been asked by a couple of readers about support for 3D
hardware support in the various X servers. I'm going to contact the
X Server vendors (Xi Graphics, MetroLink, The XFree Project)
as well as Brian Paul (the MesaGL author)
and see what they have to say. If you are connected with these folks
and have some insight I'd love to hear what you have to say. Please
email me if you know if such
support is forthcoming and I'll include it in an upcoming Graphics
Muse column.
Disclaimer:
Before I get too far into this I should note that any of the news items I
post in this section are just that - news. Either I happened to run
across
them via some mailing list I was on, via some Usenet newsgroup, or via
email from someone. I'm not necessarily endorsing these products (some of
which may be commercial), I'm just letting you know I'd heard about
them in the past month.
Frame grabber device driver for the ImageNation Cortex I video
capture card - Version 1.1
This adapter is an 512 by 486 resolution 8bit gray
scale video capture card. The device can provide data in
pgm file format or raw image data.
FTP site:
ftp://sunsite.unc.edu/pub/Linux/
apps/video/cortex.drv.1.1.tgz
Web Site:
http://www.cs.ubc.ca/spider/
jennings/cortex.drv.1.1.tgz
3DOM - a new 3D modeller project using OpenGL for Linux
3DOM is a 3D Modeler for Unix (using HP StarBase or OpenGL/Mesa)
that is free for
non-commercial use. Source code is available. Binaries for
Linux/Intel, SGI, Sparc Solaris and HP-UX are also availalbe.
cwis/research/graphics/3DOM/
Pixcon/Anitroll R1.04
I found this in my /tmp directory while getting ready for this months
column. I couldn't find a reference to it in any other Muse columns so I
guess I must have just misplaced it while preparing for an earlier issue.
Hopefully, this isn't too out of date.
Pixcon & Anitroll is a freely available 3D
rendering and animation package, complete with source.
Pixcon is a 3D renderer that creates high quality images by using a
combination of 11 rendering primitives. Anitroll is a forward kinematic
heirarchical based animation system that has some support for some
non-kinematic based animation (such as flock of birds, and autonomous
cameras). These tools are based upon the Graph library which is full
of those neat rendering and animation algorithms that those 3D faqs
keep mentioning. It also implements
some rendering techniques that were presented at Siggraph 96 by Ken
Musgrave and was used to generate an animation for Siggraph '95.
New features since version 1.03:
The Pixcon & Anitroll home page is at:
http://www.radix.net/~dunbar/index.html
Comments can be emailed to
dunbar@saltmine.radix.net
Pixcon is available either through the above web site or at Sunsite.
It is currently under:
/pub/Linux/Incoming/pixcon104.tgz
and will be moved to:
/pub/Linux/apps/graphics/pixcon104.tgz
NOTE: there is a file pixcon1.04.tgz in those directories,
but it's corrupted. Be sure to get the correct files.
ELECTROGIG 3DGO
 
ELECTROGIG
is a software company specialiced in
3D solid modeling, visualization and animation software.
The latest version of 3DGO (version 3.2), a modeling
animation and raytracing package, is now available for
the Intel Linux platform. A beta version is also available for the
MkLinux platform.
Take a look at the benchmarks for Linux
on the intel platform:
http://www.gig.nl/products/prodbench.html.
 
3DGO was originally developed for the SGI platform and
is available for the SGI, SUN and HP platforms.
 
For more comprehensive information about 3DGO, check out the
WWW-site:
http://www.gig.nl/products/prodinfo.html.
 
You can download a demo-version of 3DGO for linux, this version
has all functionality, except the save functions. Go to our
download area:
ftp://ftp.gig.nl/demo/.
Please Read the .txt files before downloading.
INFO: info@gig.nl
EZWGL v1.2, the EZ widget and graphics library.
EZWGL is a C library written on top of Xlib. It has been developed on
a Linux system and has been tested on the following platforms:
SunOS 4.1.4, OSF1 V3.2 Alpha, IRIX 5.3 Linux 1.2 and Linux 2.0.23.
It should work on all Unices with X11R6.
This release is the first one that comes with a complete postscript
manual.
xfpovray v1.2b
A new release of xfpovray, the graphical interface to
POV-Ray, has been released by
Robert S. Mallozzi.
xfpovray v1.2b requires the XForms library
and supports most of the numerous options of POV-Ray. You can view
an image of the interface and get the source code from
http://cspar.uah.edu/~mallozzir/
There is a link there to the XForms home page if you don't yet have this
library installed.
libgr V2.0.12
A new version of libgr, version 2.0.12, is now available from
ftp.ctd.comsat.com:/
pub/linux/ELF/libgr-2.0.12.tar.gz.
libgr is a collection of graphics
libraries, which includes fbm, jpeg,
pbm, pgm, ppm, pnm, png, tiff, rle.
EPSCAN - scanner driver for EPSON ES-1200C/GT-9000 scanners
EPSCAN is a scanner driver for EPSON ES-1200C/GT-9000 scanners. It
includes a driver and a nice Qt-based X frontend.
It allows previewing, and
selecting a region of an image to be scanned, as well as changing
scanner settings. It only supports scanners attached to a SCSI port,
not to the parallel port.
ftp://sunsite.unc.edu/pub/Linux/Incoming/epscan-0.1.tar.gz.
RPM versions of the binary and source are available from
ftp://ftp.redhat.com/pub/Incoming/epscan-0.1-1.src.rpm
ftp://ftp.redhat.com/pub/Incoming/epscan-0.1-1.i386.rpm.
They're intended destinations are
ftp://ftp.redhat.com/pub/contrib/epscan-0.1-1.src.rpm
ftp://ftp.redhat.com/pub/contrib/epscan-0.1-1.i386.rpm.
and
ftp://sunsite.unc.edu/pub/Linux/apps/graphics/scanners/epscan-0.1.tar.gz
Author: Adam P. Jenkins
<ajenkins@cs.umass.edu>
Inlab-Scheme Release 4
Inlab-Scheme Release 4 is now available for Linux/386 (2.X kernel,
ELF binary) and FreeBSD.
Inlab-Scheme is an independent implementation of the algorithmic
language Scheme as defined by the R4RS and the IEEE Standard
1178. In addition to the language core Inlab-Scheme has support
for bitmap/greymap processing of several kinds. Inlab-Scheme can
be used as a general tool for image processing, OCR or specialized
optical object recognition.
Inlab-Scheme Release 4 reads and writes multipage tiff/G4, XBM and
PNG graphic file formats. Inlab-Scheme Release 4 has built in
converters for various patent image file formats (USAPat,
PATENTIMAGES and ESPACE).
Inlab-Scheme is distributed at
http://www.munich.net/inlab/scheme,
where additional information about the current state of the project,
supported platforms, current license fees and more is available.
The Linux Game SDK Project
The new WWW page for the Linux Game SDK is at
http://www.ultranet.com/~bjhall/GSDK/.
The Linux GSDK Project is a new project which aims to make
a consistent and easy to use set of libraries to ease game
developers (professional or not) to make first class games
under the Linux OS. The GSDK will provide libraries for 2D and
3D graphics, advanced sound, networked games and input devices.
It should also improve the development of multimedia applications for
Linux. See the Web site for more informations.
The GSDK mailing list has moved from linux-gsdk@endirect.qc.ca to
linux-gsdk@mail.wustl.edu.
Additionnal lists have been created for the various teams.
WebMagick Image Web Generator
 
WebMagick is a package which supports making image collections available
on the Web. It recurses through directory trees, building HTML pages,
imagemap files, and client-side/server-side maps to allow the user to
navigate through collections of thumbnail images (somewhat similar to
xv's Visual Schnauzer) and select the image to view with a mouse click.
 
WebMagick is based on the "PerlMagick" ImageMagick PERL extension rather
than external ImageMagick utilities (as its predecessor "Gifmap" is). This
alone is good for at least a 40% performance increase on small images.
WebMagick supports smart caching of thumbnails to speed montage generation
on average size images. After a first pass at "normal" speed, successive
passes (upon adding or deleting images) are 5X to 10X faster due to the
caching.
 
WebMagick supports a very important new feature in its caching subsystem:
it can create and share a thumbnail cache with John Bradley's 'xv' program.
This means that if you tell 'xv' to do an update, WebMagick montages will
benefit and you can run WebMagick as a batch job to update xv's thumbnails
without having to wait for 'xv' to do its thumbnail reduction (and get a
browsable web besides!).
 
WebMagick requires the ImageMagick (3.8.4 or later) and PerlMagick (1.0 or
later) packages as well as a recent version of PERL 5.
Primary-site:
http://www.cyberramp.net/~bfriesen/webmagick/dist/webmagick-1.17.tar.gz
Alternate-site:
ftp.wizards.dupont.com/pub/ImageMagick/perl/webmagick-1.17.tar.gz
Perl Language Home Page:
http://www.perl.com/perl/index.html
ImageMagick:
http://www.wizards.dupont.com/cristy/ImageMagick.html
PerlMagick:
http://www.wizards.dupont.com/cristy/www/perl.html
Author:
Bob Friesenhahn (bfriesen@simple.dallas.tx.us)
SIMLIB IG - Commercial library
 
SIMLIB IG a C library which enables communication with Evans &
Sutherland graphics Supercomputers (so called image generators).
It enables the user to communicate with Evans & Sutherland image
generators (Liberty and ESIG Systems) using a very efficient raw
Ethernet protocol. There is no need for using opcodes, since
SIMLIB IG provides an API to the functionality of the image
generators.
 
Documentation comes printed in English, and source code
examples are provided on the distribution media.
The software is also available for SGI and NT systems.
SIMLIB IG for all other OS is $5000 (US)
Technologiezentrum Innsbruck
AUSTRIA/EUROPE
mtekscan - Linux driver for MicroTek ScanMaker SCSI scanners
mtekscan is a Linux driver for MicroTek
ScanMaker (and compatible) SCSI scanners. Originally developed for the
ScanMaker E6, it is (so far) known to also work with the ScanMaker
II/IIXE/IIHR/III/E3/35t models, as well as with the Adara ImageStar I,
Genius ColorPage-SP2 and Primax Deskscan Color.
The new version of mtekscan is v0.2. It's still in beta testing,
but all major options should work without problems. Besides some
small bugfixes and minor improvements, the new version contains a
couple of new features, most notably:
mtekscan v0.2 is available as mtekscan-0.2.tar.gz from the
Fast Forward ftp-server:
ftp://fb4-1112.uni-muenster.de/pub/ffwd/
or from sunsite:
ftp://sunsite.unc.edu/
pub/Linux/apps/graphics/scanners/
PNG binaries for Netpbm tools now available
Linux binaries for pnmtopng, pngtopnm, and gif2png are available at:
http://www.universe.digex.net/~dave/files/pngstuff.tgz
If you have trouble dowloading that, see
http://www.universe.digex.net/~dave/files/index.html
for helpful instructions.
 
PNG is the image format that renders GIF obsolete.
For details on that, you can visit the PNG home page at:
http://www.wco.com/~png/.
 
The only shared libraries you need are libc and libm; all of the
others are linked statically.
The versions of libraries used to build the programs are those
that were publicly available as of 1997-04-06:
TN-Image Version 2.5.0
TN-Image is:
It includes a 123-page manual, tutorials, and on-line help.
The Unix version is highly customizable with regard to fonts,
colors, etc.
Contact and archive information:
Did You Know?
...that there is a converter available to turn POV-Ray heightfields
into RenderMan compliant RIB files for use with BMRT? Florian
Hars writes:
I've worked on my code, now it uses libpgm and has all the necessary
vector routines included, it is on my page (with some comparisions of
rendering time and memory consumption):
http://www.math.uni-hamburg.de/home/hars/rman/height.html
Florian also has some pages of general POV vs. RenderMan comparisons:
http://www.math.uni-hamburg.de/home/hars/rman/rm_vs_pov.html
% giftopnm color_cube_colors.gif" > color_cube.ppm
% tgatoppm image.tga | ppmquant -m color_cube.ppm -fs | \
ppmtogif -interlace -transparent rgb:ff/ff/ff > image.gif
http://www.3dcafe.com/anatomy.htm
P2
# kilimajaro.pgm
15 10
59
10 15 18 20 21 22 23 23 21 20 19 18 17 16 15
11 15 19 22 27 30 30 30 29 28 25 20 19 18 17
13 15 19 21 28 38 36 40 40 35 30 24 20 19 18
15 16 18 20 29 39 37 44 59 44 38 30 22 19 18
15 16 18 20 28 30 30 40 50 46 51 48 28 20 19
15 15 16 17 18 19 20 24 30 35 37 37 30 20 19
15 15 14 15 16 17 18 19 22 29 30 29 27 20 19
15 15 14 13 15 16 15 17 18 20 22 20 20 20 18
15 14 13 11 12 12 12 13 14 15 17 15 15 15 14
14 11 10 9 9 10 10 10 9 10 13 12 11 11 11
Use it with scale <15,1.18,10> to get an to-scale image and with a
larger y-scale if you want to see something. The earth is incredibly
flat.
ObjectBegin 2
SolidBegin "primitive"
TransformBegin
Translate -1 0 0
Rotate -90 0 1 0
Patch "bilinear" "P"
[ -1 -1 0 1 -1 0
-1 1 0 1 1 0 ]
TransformEnd
... etc.
ObjectEnd
 
Transformations just aren't allowed inside object definitions.
Remember that object instances inherit the entire graphics state
that's active when they are instanced -- including transformations.
So all primitives within the instanced object get the very same
transformation. If they're all bilinears like you have them, that
means that they will all end up on top of one another.
 
For this reason and others, I urge everybody to not use instanced
objects at all for any RenderMan compliant renderer. They're quite
useless as described in the RenderMan 3.1 spec.
Yes, I know that RenderMan Companion has an example that does
exactly what I said is illegal. The example is wrong, and will not
work with either PRMan or BMRT.
 
Solid (CSG) operations are meant only to operate on solids.
A solid is a boundary representation which divides space into three
distinct loci: (a) the boundary itself, which has finite surface area,
(b) a (possibly disconnected) region of finite volume (the "inside"),
and (c) a connected region of infinite volume (the "outside").
You can't subtract a box from a bilinear patch, since a bilinear patch
isn't a solid to begin with.
 
If you want a flat surface with a square hole, there are two methods
that I'd recommend: (a) simply use several bilinears (4 to be exact)
for the surface, like this:
+-----------------------+
| #1 |
| |
+======+---------+======+
| | | |
| #2 | (hole) | #3 |
| | | |
+======+---------+======+
| #4 |
| |
+-----------------------+
Or, (B) if you really want to be fancy, use a single flat order 2
NURBS patch with an order 2 trim curve to cut out a hole.
Correcting for display gamma
This past 2 months I've been hard at work on an entry for this
round of the IRTC, the Internet Ray Tracing Competition. In
previous rounds I had submitted entries using POV-Ray, but for
this round I switched to BMRT, mostly so I could learn the RenderMan
API and how to write shaders using the RenderMan shading language.
This months main article is the second of a three part series on
BMRT. The BMRT package is written by Larry Gritz, and Larry was
gracious enough to offer some wonderful critiques and tips on
my image.
During out email correspondence, Larry noticed I had overlit my
scenes quite badly. While we tried to figure out what was causing
this (it turned out to be a misuse of some parameters to some
spotlights I was using) he asked if I had gamma corrected for my
display. Gamma correction is a big issue in computer graphics, one
that is often overlooked by novices. I'd heard and read quite a
bit about gamma correction but had never really attempted to determine
how to adjust the gamma for my display. Larry offered an explanation,
a quick way to test the gamma on my system, and a tip for
adjusting for gamma correction directly in the BMRT renderer, rendrib.
I thought this would be a great thing to share with my readers, so
here it is.
Rendrib produces linear pixels for its output -- i.e. a pixel with
value 200 represents twice as much light as a pixel of value 100.
Thus, it's expected that your display will be twice as bright
(photometrically, not necessarily perceptually) on a pixel of 200
than one of 100.
This sort of display only really happens if you correct for gamma,
the nonlinearity of your monitor. In order to check this,
take a look at the following chart. Display the chart as you'd
view any image. You'll notice that if you squint, the apparent
brightness of the left side will match some particular number on
the right. This is your gamma correction factor that must be
applied to the image to get linear response on your particular
monitor.
If your display program uses Mesa (as rendrib's framebuffer display
does), you can set an environment variable, MESA_GAMMA, to this value
and it will transparently do the correction as it writes pixels to the
screen. Most display programs let you correct gamma when you view
an image, though I've had trouble getting xv to do it without messing
up the colors in a weird way.
Another alternative is to put the following line in your RIB file:
Exposure 1 <gamma>
More Musings...
My Entry in the March/April IRTC
- a case study in learning to
use RenderMan and BMRT
where gamma was what you measured with the chart.
This will cause rendrib to pre-correct the output pixels for the
gamma of your display.
I think it's important to gamma correct so that at least you're
viewing the images the way that rendrib "expects" them to appear.
It can't know about the nonlinearities of your CRT without you
telling it.
Larry has more on the gamma issue on his own pages. You can find
it at
http://www.seas.gwu.edu/student/gritz/gamma.html.
He also asked me to mention that he got this chart from
Greg Ward, but we didn't have any contact information for him.
Hopefully he doesn't mind our using it.
Readers should note that the image displayed in this article may
not provide accurate information for adjusting gamma since your
browser may dither the colors in a way which changes what the
actual value should be. Also, this image is a JPEG version of
the original TIFF image Larry supplied. Its possible the conversion
also changed the image. If you're interested in trying this out you
should grab
the original TIFF image
(300x832).
The following links are just starting points for finding more information
about computer graphics and multimedia in general for Linux systems. If
you have some application specific information for me, I'll add them to my
other pages or you can contact the maintainer of some other web site. I'll
consider adding other general references here, but application or site
specific information needs to go into one of the following general
references and not listed here.
Unix Graphics Utilities
Linux Multimedia Page
The IRTC-L discussion list
comp.graphics.rendering.raytracing
comp.graphics.rendering.renderman
comp.os.linux.announce
Future Directions
Next month:
Let me know what you'd like to hear about!
Copyright © 1997, Michael J. Hammel
Published in Issue 17 of the Linux Gazette, May 1997
More...
© 1997 Michael J. Hammel
My Entry in the March/April IRTC - a case study in learning to
use RenderMan and BMRT
One reason I took so long to get down to writing the Muse column this month
was that I was hard at work on an entry in the IRTC, the Internet Relay
Tracing Competition, which I help administer. I've been an active
participant in the IRTC since its restart back in May 1996 but have only
actually entered one competition. So this round had a special meaning for
me. I don't often have the time to work on entries unless something else
suffers. In this case, it was last months Muse column. To be honest,
however, I was also using this entry to learn more about RenderMan, BMRT
and in particular, the RenderMan Shading Language. Nothing is quite such a
a teacher as experience. And my entry in the IRTC was a wonderful teacher.
Below I've included the text file which accompanies my entry in the
IRTC. All entries must have this file. It describes the who/what/how and
so forth relating to an entry. I'll let the text file describe what I did,
who helped me do it, and some of the issues I encountered. I hope you find
this useful information.
IMAGE DESCRIPTION:
EMAIL:
mjhammel@csn.net
NAME:
Michael J. Hammel
TOPIC:
school
TITLE:
Post Detention
RENDERER USED:
BMRT
TOOLS USED:
Linux, AC3D, BMRT, Gimp 0.99.8,
wc2pov, xcalc, xv
RENDER TIME:
about 4 hours
HARDWARE USED:
Cyrix P166 (133Mhz)
based Linux system, 64M memory
Pretty simplistic, really. Its a school room, just after detention has let
out. You can tell that detention has let out by the writing on the chalk
board and the time displayed by the clock on the wall. The sun is starting
to get low outside, which causes the types of shadows you can see on the
bookshelf. All the students who were in detention are required to read the
latest New York Times bestseller titled "Post Detention". Its written by
some author who is rumored to do 3D graphics on the side. You can see the
book on the desk in the lower right corner of the image.
I used this image to learn to use RenderMan and BMRT. I find I like these
tools a bit better than POV-Ray, mostly because I can write C code to
create models if I want (although I didn't for this particular scene) and
the shader language allows a bit more control than POV's. I still have
much to learn to make real use of these features, however.
I started with some canned models from 3DCafe: a chair, a couple of
bookcases, and some books. I had to convert the chair from 3DS to DXF so I
could then import it into AC3D. Once I had it in the modeler, I broke the
chair into two pieces - the arm and the rest of the chair. I did this so I
could texture them seperately (note that in the 3DS format these pieces may
have already been seperate, but after the conversion to DXF they were a
single entity). I also sized the chair to be a common unit size and
centered it on the origin. This unit size was used on all models so that
the real sizing and positioning could be done in the RIB file.
The book case only needed resizing but the books had to be broken into a
cover and "text pages". The latter are a single entity that were textured
with a matte finish. The covers were textured individually. All the books
are basically the same book sized differently and placed in the bookcase
from within AC3D. This provided relative positioning for the books in the
bookcase, after which any other translations would maintain that relative
positioning. Books on the chairs or floor were done similarly.
The walls are simple polygons or bilinear patches. The windowed wall
turned out to show problems in the way AC3D does polygons for RIB output.
I had to convert this wall from polygons to a set of bilinear patches (see
the May issue of the Graphics Muse for
how to make a hole in a wall using bilinear
patches)
in order for the texture to be applied evenly over all of
the patches. This problem also showed up when trying to apply texture maps
to the chalk board. It apparently has to do with vertex ordering. I had
to change the chalkboard to a bilinear patch too. I may have to write my
own models for anything that uses texture maps in the future instead of
using AC3D. To his credit, I haven't told Andy (AC3D's author) about this
yet and he's been very good about addressing any problems I've found in the
past.
An important aspect of this image is the tremendous help I got from Larry
Gritz, the author of BMRT. He offered some special shaders, although I
only used one of them (the one for the ceiling tiles). The biggest help
was very constructive criticism and tips for things too look for in my
code. For example, he pointed out that I was probably using the parameters
to spotlights incorrectly. I was, and it was causing my scene to be
very
overlit (spotlights use coneangle in radians, and I had specified them in
degrees). This one change made a dramatic improvement in the overall
image.
All the shaders used, except for the ceiling tiles and the papers on the
desks, are from the archive of shaders from Guido Quaroni. This archive
includes shaders from the RenderMan Companion by Steve Upstill, from
Texturing and Modeling by Ebert, Musgrave, et al, Larry Gritz, and various
other places. Its quite a useful resource for novices just getting started
with shaders and RenderMan. The papers on the desks are textured with a
shader I wrote that creates the horizonatl and vertical lines. It also
puts in 3 hole punches, but thats not obvious from the image. This shader
is the only one I included in the source [Ed. The IRTC allows the source
code for scenes to be uploaded along with the JPEG images and text file].
The chairs are textured with a
rusty metallic shader and a displacement shader for the dents.
Displacement shaders are cool because they actually move points on the
surface (unlike bump maps which just change the normals for those points).
The arm surfaces are textured with a wood shader that I made a minor change
to (to allow better control of the direction of the wood grain) and a
displacement shader that caused the bumpiness, scratches, and chips in the
wood. This latter item could have been better, but I was running out of
time.
The chalkboard is an image map created in the 0.99.8 developers release of
the Gimp. This is a very cool tool for Linux and Unix platforms, very
similar to Photoshop (but apparently better, according to people who've
used both - I've never used Photoshop myself).
The image out the window is an image map on a simple plane angled away from
the window. The window panes are dented using a displacement map. We had
windows with bumps in them in High School and thats the effect I was going
for here. Its pretty close and as an added benefit it prevents the image
outside the window from being too washed out.
The globe on the book shelf is one of those ones that is suspended by
magnets. The globe has a displacement map on it as well, which is why,
if you look real close, the lighting on it is not smooth where it moves
into shadow. The globe and its base were completely modeled in AC3D. It
was quick and very easy to do. All the items on the booksshelf by the
window are in a single model file, but exported as individual objects so
they could be shaded properly. The same is true for the bookcase under the
clock.
It was fun. This is certainly the best 3D image I've done to date. Its
also the first one of something recognizable (as opposed to space ships and
planets no one has ever really seen).
NOTE:
One thing I forgot to mention in my original text file for my entry is that
I had to edit all the exported RIB files that I created with AC3D to remove
the bits of RIB that made each file an independently renderable file. By
default AC3D generates a complete scene, one that can be passed to rendrib
(the BMRT renderer) directly to render a scene. But what I needed was for
the files generated to be only partially complete scenes, without the
camera or lighting and so forth. In this way I could use these files in
RIB ReadArchive statements, similar to #include files for POV. Considering
the number of objects I exported with AC3D, that turned out to be quite a
bit of hand editing. I sent email to Andy Colebourne, the author of AC3D,
and he's looking into making it possible to output partial RIBs for use as
ReadArchive include files.
© 1997 by Michael J. Hammel
More...
BMRT
© 1996 Michael J. Hammel
1. A quick review
Before we get started on shaders, lets take a quick look back at RIB files.
RIB files are ASCII text files which describe a 3D scene to a RenderMan
compliant renderer such as BMRT. A RIB file contains descriptions of
objects - their size, position in 3D space, the lights that illuminate them
and so forth. Objects have surfaces that can be colored and textured,
allowing for reflectivity, opacity (or conversely, transparency),
bumpiness, and various other aspects.
An object is instanced inside AttributeBegin/AttributeEnd requests (or
procedures in the C binding). This instancing causes the current graphics
state to be saved so that any changes made to the graphics state (via the
coloring and texturing of the object instance) inside the
AttributeBegin/AttributeEnd request will not affect future objects. The
current graphics state can be modified, and objects colored and textured,
with special procedures called shaders.
Note:
Keep in mind that this is not a full fledged tutorial and I won't be
covering every aspect of shaders use and design. Detailed information can
be found in the texts listed in the bibliography at the end of this
article.
2. What is a shader?
In the past, I've often used the terms shading and texturing
interchangeably. Darwyn Peachy, in his
Building Procedural Textures
chapter in the text Texturing and Modeling: A Procedural Approach,
says that these two concepts are actually separate processes:
Shading is the process of calculating the color of a pixel
from user-specified surface properties and the shading model.
Texturing is a method of varying the surface properties from
point to point in order to give the appearance of surface
detail that is not actually present in the geometry of the
surface.
[1]
A shader is a procedure called by the renderer to apply colors and textures
to an object. This can include the surface of objects like block or spheres,
the internal space of a solid object, or even the space between objects
(the atmosphere). Although based on Peachy's description would imply that
shaders only affect the coloring of surfaces (or atmosphere, etc), shaders
handle both shading and texturing in the RenderMan environment.
3. Compiling shaders
RIB files use filenames with a suffix of ".rib". Similarly, shader files
use the suffix ".sl" for the shader source code. Unlike RIB files, however,
shader files cannot be used by the renderer directly in their source
format. They must be compiled by a shader compiler. In the BMRT package
the shader compiler is called slc.
Compiling shaders is fairly straightforward - simply use the slc program
and provide the name of the shader source file. For example, if you have a
shader source file named myshader.sl you would compile it with the
following command:
slc myshader.sl
You must provide the ".sl" suffix - the shader source file cannot be
specified using the base portion of the filename alone.
When the compiler has finished it will have created the compiled shader in
a file named myshader.so
in the current directory. A quick examination of
this file shows it to be an ASCII text file as well, but the format is
specific for the renderer in order for it to implement its graphics state
stack.
Note: the filename extension of ".so" used by BMRT (which is different
than the one used by PRMan) does not signify a binary object file, like
shared library object files. The file is an ASCII text file. Larry says
he's considering changing to a different extension in the future to avoid
confusion with shared object files.
Note that in the RIB file (or similarly when using the C binding)
the call to the shader procedure is done in the following manner:
AttributeBegin
Color [0.9 0.6 0.6]
Surface "myshader"
ReadArchive "object.rib"
AttributeEnd
This example uses a surface shader (we'll talk about shader types in a
moment). The name in double quotes is the name of the shader procedure
which is not necessarily the name of the shader source file.
Since shaders are procedures they
have procedure names. In the above example the procedure name is
myshader. This happens to the be same as the base portion (without
the suffix) of the shader source filename. The shader compiler doesn't
concern itself with the name of the source file, however, other than to
know which file to compile. The output filename used for the .so file is
the name of the procedure. So if you name your procedure differently than
the source file you'll get a differently named compiled .so file. Although
this isn't necessarily bad, it does make it a little hard to keep track of
your shaders. In any case, the name of the procedure is the name used in
the RIB (or C binding) when calling the shader. In the above example,
"myshader" is the name of the procedure, not the name of the source file.
4. Types of shaders
According to the RenderMan Companion [2]
The RenderMan Interface specifies six types of shaders, distinguished
by the inputs they use and the kinds of output they produce.
The text then goes on to describe the following shader types:
Most of these can only have one instance of the shader type in the graphics
state at any one time. For example, there can only be one surface shader
in use for any object or objects at a time. The exception to this are
light shaders, which may have many instances at any one time, some of which
may not be actually turned on for some objects.
Light sources in the RenderMan Shading Language are provided a position and
direction and return the color of the light originating from that light and
striking the current surface point. The RenderMan specification provides
for a set of default light shaders that are very useful and probably cover
the most common lighting configurations an average user might encounter.
These default shaders include ambient light (the same amount of light
thrown in all directions), distant lights (such as the Sun), point lights,
spot lights, and area lights. All light sources have an intensity that
defines how bright the light shines. Lights can be made to cast shadows or not
cast shadows. The more lights that cast shadows you have in a scene the
longer it is likely to take to render the final image. During scene design
and testing its often advantagous to keep shadows turned off for most lights.
When the scene is ready for its final rendering turn the shadows back on.
Ambient light can be used to brighten up a generally dark image but the
effect is "fake" and can cause an image to be washed out, losing its
realism. Ambient light should be kept small for any scene, say with an
intensity of no more than 0.03. Distant lights provide a light that shines
in one direction with all rays being parallel. The Sun is the most common
example of a distant light source. Stars are also considered distant
lights. If a scene is to be lit by sunlight it is often considered a good
idea to have distant lights be the only lights to cast shadows. Distant
lights do not have position, only direction.
Spot lights are the familiar lights which sit at a particular location in
space and shine in one generalized direction covering an area specified by
a cone whose tip is the spot light. A spot lights intensity falls off
exponentially with the angle from the centerline of the cone. The angle is
specified in radians, not degress as with POV-Ray. Specifying the
angle in degrees can have the effect of severly over lighting the area
covered by the spot light. Point lights also fall off in intensity, but do
so with distance from the lights location. A point light shines in all
directions at once so does not contain direction but does have position.
Area lights are series of point lights that take on the shape of an object
to which they are attached. In this way a the harshness of the shadows
cast by a point light can be lessened by creating a larger surface of
emitted light. I was not able to learn much about area lights so can't
really go into detail on how to use them here.
Most light source shaders use one of
two illumination functions: illuminate() and solar(). Both provides ways
of integrating light sources on a surface over a finite cone. illuminate()
allows for the specification of position for the light source, while
solar() is used for light sources that are considered very distant, like
the Sun or stars. I consider the writing of light source shaders to be a
bit of an advanced topic since the use of the default light source shaders
should be sufficient for the novice user to which this article is aimed.
Readers should consult The RenderMan Companion and The RenderMan
Specification for details on the use of the default shaders.
Surface shaders are one of the two types of shaders novice users will make
use of most often (the other is displacement shaders). Surface shaders are
used to determine the color of light reflected by a given surface point
in a particular direction. Surface shaders are used to create wood
grains or the colors of an eyeball. They also define the opacity of a
surface, ie the amount of light that can pass through a point (the points
transparency). A point that is totally opaque allows no light to pass
through it, while a point that is completely transparent reflects no light.
The majority of the examples which follow will cover surface shaders. One
will be a displacement shader.
A volume shader affects light traveling to towards the camera as it passes
though and around objects in a scene. Interior volume shaders determine
the effect on the light as it passes through an object. Exterior volume
shaders affect the light in the "empty space" around an object.
Atmospheric shaders handle the space between objects. Exterior
and interior volume
shaders differ from atmospheric shaders in that the latter operate on all
rays originating from the camera (remember that ray tracing traces the
lights ray in reverse from nature - from camera to light source).
Exterior and interior shaders work only on secondary rays, those rays
spawned by the trace() function in shaders.
Atmospheric shaders are used
for things like fog and mist. Volume shaders are a slightly more advanced
topic which I'll try to cover in a future article.
The texture of an object can vary in many ways, from very smooth to very
bumpy, from smooth bumps to jagged edges. With ordinary surface shaders a
texture can be simulated with the use of a bump map. Bump maps
perturb the normal of a point on the surface of an object so that the point
appears to be raised, lowered, or otherwised moved from its real location.
A bump map describes the variations in a surfaces orientation.
Unfortunately, this is only a trick and the surface point is not really
moved. For some surfaces this trick works well when viewed from the proper
angle. But when seen edge on the surface variations disapper - the edge is
smooth. A common example is an orange. With a bump map applied the orange
appears to be pitted over its surface. The edge of the sphere, however, is
smooth and the pitting effect is lost. This is where displacement shaders
come in.
In The RenderMan Interface Specification[3]
it says
The displacement shader environment is very similar to a surface shader,
except that it only has access to the geometric surface parameters.
[A displacement shader] computes a new P [point] and/or a new N
[normal for that point].
A displacement shader operates across a surface, modifying the physical
location of each point. These modifications are generally minor and of a
type that would be much more difficult (and computationally expensive) to
specify individually. It might be difficult to appreciate this feature
until you've seen what it can do.
Plate 9 in [4] shows an ordinary cylinder modified with
the threads() displacement shader to create the threads on the base
of a lightbulb. Figures 1-3 shows a similar (but less sophisticated)
example.
Without the use of the displacement shader, each thread
would have to be made with one ore more individual objects. Even if the
computational expense for the added objects were small, the effort required
to model these objects correctly would still be significant. Displacement
shaders offer procedural control over the shape of an object.
An ordinary cylinder
Note that in this case the renderer attributes have not been
turned on. The edges of the cylinder are flat, despite the
apparent non-flat surface.
Same cylinder with true displacements
In this image the renderer attributes have been turned on. The
edges of the cylinder reflect the new shape of the cylinder.
Figure 1
Figure 2
Figure 3
An important point to remember when using displacement shaders with
BMRT is that, by default, displacements are not turned on. Even if
a displacement shader is called the points on the surface only have
their normals modified by the shader. In order to do the "true
displacement", two renderer attribute options must be set:
Attribute "render" "truedisplacement" 1
Attribute "displacementbound" "coordinatesystem"
"object" "sphere" 2
The first of these turns on the true displacement attribute so that
displacement shaders actually modify the position of a point on the
surface. The second specifies how much the bounding box around the
object should grow in order to enclose the modified points.
How this works is that the attribute tells the renderer how much the
bounding box is likely to grow in object space. The renderer can't
no before hand how much a shader might modify a surface, so this statement
provides a maximum to help the renderer with bounding boxes around
displacement mapped objects. Remember that bounding boxes are used help
speed up ray-object hit tests by the renderer. Note that you can compute
the possible change caused by the displacement in some other space, such as
world or camera. Use whatever is convenient. The "sphere" tag lets the
renderer know that the bounding box will grow in all directions evenly.
Currently BMRT only supports growth in this manner, so no other values
should be used here.
BMRT doesn't support Transformation Shaders (neither does Pixar's PRMan
apparently). Apparently transformation shaders are supposed to operate on
geometric coordinates to apply "non-linear geometric transformations".
According to [5]
The purpose of a transformation shader is to modify a coordinate system.
It is used to deform the geometry of a scene without respect to any
particular surface. This differs from a displacement shader because the
displacement shader operates on a point-by-point basis for a given surface.
Transformation shaders modify the current transform, which means they
can affect all the objects in a scene.
Imager shaders appear to operate on the colors of output pixels which to me
means the shader allows for color correction or other manipulation after a
pixels color has been computed but prior to the final pixel output to file
or display. This seems simple enough to understand, but why you'd use them
I'm not quite sure. Larry says that BMRT supports Imager shaders but PRMan
does not. However, he suggests the functionality provided is probably
better suited to post-processing tools, such as XV, ImageMagick or the Gimp.
5. Shader language syntax
So what does a shader file look like? They are very similar in format to a
C procedure, with a few important differences. The following is a very
simplistic surface shader:
surface matte (
float Ka = 1;
float Kd = 1;
)
{
point Nf;
/*
* Calculate the normal which is facing the
* direction that points towards the camera.
*/
Nf = faceforward (normalize(N),I);
Oi = Os;
Ci = Os * Cs * (Ka * ambient() + Kd * diffuse(Nf));
}
This is the matte surface shader provided in the BMRT distribution. The
matte surface shader happens to be one of a number of required shaders that
The RenderMan Interface Specification says a RenderMan compliant
renderer must provide.
The first thing to notice is the procedure type and name. In this case the
shader is a surface shader and its name is "matte". When this code is
compiled by slc it will produce a shader called "matte" in a file called
"matte.so". Procedure names can be any name that is not a reserved RIB
statement. Procedure names may contain letters, numbers and underscores.
They may not contain spaces.
There are a number of different kinds of variables that are used with
shaders: Instance variables, global variables, and local variables.
Instance variables are the variables used as parameters to the shader.
When calling a shader these variables are declared (if they have not
already been declared) and assigned a value to be used for that instance of
the shader. For example, the matte shader provides two parameters that can
have appropriate values specified when the shader is instanced within the
RIB file. Lets say we have a sphere for which we will shade using the
matte shader. We would specify the instance variables like so:
AttributeBegin
Declare "Kd" "float"
Declare "Ka" "float"
Surface "matte" "Kd" 0.5 "Ka" 0.5
Sphere 1 -.5 .5 360
AttributeEnd
The values specified for Kd and Ks are the instance variables and the
renderer will use these values for this instance of the shader. Instance
variables are generally known only to the shader upon the initial call for
the current instance.
Local variables are defined within the shader itself and as such are only
known within the shader. In the example matte shader, the variable Nf is a
point variable as has meaning and value only within the scope of the shader
itself. Other shaders will not have access to the values Nf holds.
Local variables are used to hold temporary values required to compute the
values passed back to the renderer. These return values are passed back as
global variables.
Global variables have a special place in the RenderMan environment. The
only way a shader can pass values back to the renderer is through global
variables. Some of the global variables that a shader can manipulate are
the surface color (Cs), surface opacity (Os), the normal vector for the
current point (N) and the incident ray opacity
(Oi). Setting these values within the shader affects how the renderer
colors surface points for the object which is being shaded. The complete
list of global variables that a particular shader type can read or modify
is listed in tables in the RenderMan Interface Specification
[6].
Global variables are global in the sense that they pass values between the
shader and the renderer for the current surface point, but they cannot be
used to pass values from one objects shader to another.
Shaders have access to only 4 data types: one scalar type, two vector
types, and a string type. A string can be defined and used by a shader, but
it cannot be modified. So an instance variable that passes in a string
value cannot be modified by the shader, nor can a local string variable be
modified once it has been defined.
The scaler type used by shaders is called a float type. Shaders must use
float variables even for integer calculations. The point type is a
a 3 element array of float values which describe a point in some space.
By default the point is in world space in BMRT (PRMan uses camera
space by default), but it is possible to
convert the point to object, world, texture or some other space within the
shader. On point can be transformed to a different space using the
transform statement. For example:
float y = ycomp(transform("object",P));
will convert the current point to object space and return the Y component
of the new point into the float variable y. The other vector type is also
a 3 element array of float values that specify a color. A color
type variable can be defined as follows:
color Cp = color (0.5, 0.5, 0.5);
Expressions in the shading language follow the same rules of precedence
that are used in the C language. The only two expressions that are new to
shaders are the Dot Product and the Cross Product. The Dot Product is
used to measure the angle between two vectors and is denoted by a period
(.). Dot Products work on point variables.
The Cross Product is often used to find the normal vector
at a point given two nonparallel vectors tangent to the surface at a given
point. The Cross Product only works on points, is denoted by a caret (^)
and returns a point value.
A shader need not be a completely self contained entity. It can call
external routines, known as functions. The RenderMan Interface
Specificatoin predefines a large number of functions that
are available to shader authors using BMRT. The following list is just
a sample of these predefined functions:
This is not a comprehensive list, but it provides a sample of the
functions available to the shader author. Many functions operate on more
than one data type (such as points or colors). Each can be used to
calculate a new color, point, or float value which can then be applied to
the current surface point.
Shaders can use their own set of functions defined locally. In fact, its
often helpful to put functions into a function library that can be included
in a shader using the #include directive. For example, the
RManNotes Web site
provides a function library called "rmannotes.sl" which contains a
pulse() function that can be used to create lines on a surface. If
we were to use this function in the matte shader example, it might look
something like this:
#include "rmannotes.sl"
surface matte (
float Ka = 1;
float Kd = 1;
)
{
point Nf;
float fuzz = 0.05
color Ol;
/*
* Calculate the normal which is facing the
* direction that points towards the camera.
*/
Nf = faceforward (normalize(N),I);
Ol = pulse(0.35, 0.65, fuzz, s);
Oi = Os*Ol;
Ci = Os * Cs * (Ka * ambient() + Kd * diffuse(Nf));
}
The actual function is defined in the rmmannotes.sl file as
#define pulse(a,b,fuzz,x) (smoothstep((a)-(fuzz),(a),(x)) - \
smoothstep((b)-(fuzz),(b),(x)))
A shader could just as easily contain the #defined value directly without
including another file, but if the function is useful shader authors may
wish to keep them in a separate library similar to rmmannotes.sl. In this
example, the variable s is the left-to-right component of the current
texture coordinate. "s" is a component of the texture space, which
we'll cover in the section on coordinate systems. "s" is a global variable
which is why it is not defined within the sample code.
The shading language provides the following statements for flow control:
All of these act just like their C counterparts.
There are number of coordinate systems used by RenderMan. Some of these I
find easy to understand by themselves, others are more difficult -
especially when used within shaders. In a shader, the surface of an object
is mapped to a 2 dimensional rectangular grid. This grid runs from
coordinates (0,0) in the upper left corner to (1,1) in the lower right
corner. The grid is overlayed on the surface, so on a rectangular patch
the mapping is obvious. On a sphere the upper corners of the grid map to
the same point on the top of the sphere. This grid is known as
parameter space and any point in this space is referred to by the
global variables u and v. For example, a point on the
surface which is in the exact center of the grid would have (u,v)
coordinates (.5, .5).
Similar to parameter space is texture space. Texture space is a
mapping of a texture map that also runs from 0 to 1, but the variables used
for texture space are s and t. By default, texture space is
equivalent to parameter space unless either vertex variables (variables
applied to vertices of primitive objects like patches or polygons) or the
TextureCoordinates statement have modified the texture space of the primitive
being shaded. Using the default then, a texture map image would have its upper
left corner mapped to the upper left corner of the parameter space grid
overlying the objects surface, and the lower right corner of the image
would be mapped to the lower right corner of the grid. The image would
therefore cover the entire object. Since the texture space does not have
to be equivalent to parameter space it would be possible to map an image to
only a portion of an object. Unfortunately, I didn't get far enough this
month to provide an example of how to do this. Maybe next month.
There are other spaces as well: world space, object space, and shader space.
How each of these affects the shading and texturing characteristics is not
completely clear to me yet. Shader space is the default space in which
shaders operate, but points in shader space can be transformed to world or
object space before being operated on. I don't know exactly what this
means or why you'd want to do it just yet
6. Format of a shader file
Shader files are fairly free form, but there are
methodologies that can be used to make writing shaders easier and the code
more understandable. In his
RManNotes [7], Stephen F. May writes
One of the most fundamental problem solving techniques is "divide and
conquer." That is, break down a complex problem into simpler parts;
solve the simpler parts; then combine those parts to
solve the original complex problem.
The basic structure of a shader is similar to a procedure in C - the shader
is declared to be a particular type (surface, displacement, and so forth)
and a set of typed parameters are given. Unlike C, however, shader
parameters are required to have default values provided. In this way a
shader may be instanced without the use of any instance variables. If any
of the parameters are specified with instance variables then the value in
the instance variable overrides the parameters default value. An
minimalist shader might look like the following:
surface null ()
{
}
In fact, this is exactly the definition of the null shader. Don't ask me
why such a shader exists. I'm sure the authors of the specification had a
reason. I just don't know what it is. Adding a few parameters, we start
to see the matte shader forming:
surface matte (
float Ka = 1;
float Kd = 1;
)
{
}
The parameters Ka and Kd have their default values provided. Note that Ka
is commonly used in the shaders in Guido Quaroni's archive of shaders to
represent a scaling factor for ambient light. Similarly, Kd is used to
scale diffuse light. These are not global variables, but they are well
known variables, much like "i", "j", and "k" are often used as counters in
C source code (a throwback to the heady days of Fortran programming).
After the declaration of the shader and its parameters comes the set of
local variables and the shader code that does the "real work". Again, we
look at the matte shader:
#include "rmannotes.sl"
surface matte (
float Ka = 1;
float Kd = 1;
)
{
point Nf;
float fuzz = 0.05
color Ol;
/*
* Calculate the normal which is facing the
* direction that points towards the camera.
*/
Nf = faceforward (normalize(N),I);
Ol = pulse(0.35, 0.65, fuzz, s);
Oi = Os*Ol;
Ci = Os * Cs * (Ka * ambient() + Kd * diffuse(Nf));
}
Nothing special here. It looks very much like your average C procedure.
Now we get into methodologies. May [8] shows us how a
layered shader's psuedo-code might look:
surface banana(...)
{
/* background (layer 0) */
surface_color = yellow-green variations;
/* layer 1 */
layer = fibers;
surface_color = composite layer on surface_color;
/* layer 2 */
layer = bruises;
surface_color = composite layer on surface_color;
/* layer 3 */
layer = bites;
surface_color = composite layer on surface_color;
/* illumination */
surface_color = illumination based on surface_color
and illum params;
/* output */
Ci = surface_color;
}
What is happening here is that the lowest level applies yellow-and green
colors to the surface, after which a second layer has fiber colors
composited (blended or overlayed) in. This continues for each of 4 defined
layers (0 through 3) plus an illumination calculation to determine the
relative brightness of the current point. Finally, the newly computed
surface color is ouput via a global variable.
Using this sort of methodology makes writing a shader much easier as well
as allowing other shader authors to debug and/or extend the shader in the
future. A shader file is therefore sort of bottom-up design, where the
bottom layers of the surface are calculated first and the topmost layers
are computed last.
7. A word about texture maps
As discussed earlier, texture maps are images mapped from 0 to 1 from left
to right and top to bottom upon a surface. Every sample in the image is
interpolated between 0 and 1. The mapping does not have to apply to the
entire surface of an object, however, and when used in conjunction with the
parameter space of the surface (the u,v coordinates) it should be possible
to map an image to a section of a surface.
Unfortunately, I wasn't able to determine exactly how to use this knowledge
for the image I submitted to the IRTC this month. Had I figured it out
in time, I could have provided text labels on the bindings of the books in
the bookcases for that scene. Hopefully, I'll figure this out in time for
the next article on BMRT and can provide an example on how to apply texture
maps to portions of surfaces.
8. Working examples
The best way to actually learn how to write a shader is to get down and
dirty in the bowels of a few examples. All the references listed in the
bibliography have much better explanations for the exaples I'm about to
describe, but these should be easy enough to follow for novices.
This example is taken verbatim from RManNotes by Stephen F. May.
The shader creates a two color cross pattern. In this example the pattern
is applied to a simple plane (a bilinear patch). Take a look at the
source code.
color surface_color, layer_color;
color surface_opac, layer_opac;
Finally, the incident rays opacity global variable is set along
with its color.
The first thing you notice is that this shader defines two local
color variables: surface_color and layer_color.
The layer_color variable is used to compute the current layers color.
The surface_color variable is used to composite the various layers
of the shader. Two other variables, surface_opacity and
layer_opacity, work similarly for the opacity of the current layer.
The first layer is a verticle stripe. The shader defines the color
for this layer and then determines the opacity for the current point
by using a function called pulse(). This is a function
provided by May in his "rmannotes.sl" function library. The pulse()
function allows the edges of the stripes in this shader to flow
smoothly from one color to another (take a look at the edges of
the stripes in the
sample image). pulse() uses the fuzz variable to determine how
fuzzy the edges will be.
Finaly, for each layer the layers color and opacity
are blended together to get the new surface color. The blend()
function is also part of rmannotes.sl and is an extension of the
RenderMan Interface's mix() function, which mixes color and opacity
values.
Figure 4
RIB Source code for this example
Oi = surface_opac;
Ci = surface_opac * surface_color;
These two values are used by the renderer to compute
pixel values in the output image.
This example is taken from the RenderMan Companion. It shows how a shader
can be used to cut out portions of a solid surface. We use the first
example as a backdrop for a sphere that is shaded with the screen() shader
from the RenderMan Companion text (the name of the shader as used here is
slightly different because it is taken from the collection of shaders from
Guido Quaroni, who changed the names of some shaders to reflect their
origins). First lets look at the sceen using the "plastic" shader
(which comes as a default shader in the BRMT distribution). Figure 5 shows
how this scene renders. The sphere is solid in this example. The
RIB code for this contains the following lines:
AttributeBegin
Color [ 1.0 0.5 0.5 ]
Surface "plastic"
Sphere 1 -1 1 360
AttributeEnd
In Figure 6 the sphere has been changed to a wireframe surface. The
only difference between this scene and Figure 5 is the surface shader used.
For Figure 6 the rib code looks like this:
AttributeBegin
Color [ 1.0 0.5 0.5 ]
Surface "RCScreen"
Sphere 1 -1 1 360
AttributeEnd
The rest of the RIBs are exactly the same. Now lets look at the
screen() shader code.
surface
RCScreen(
float Ks = .5,
Kd = .5,
Ka = .1,
roughness = .1,
density = .25,
frequency = 20;
color specularcolor = color (1,1,1) )
{
varying point Nf =
faceforward( normalize(N), I );
point V = normalize(-I);
Figure 5
RIB Source code for this example
Figure 6
RIB Source code for this example
Figure 7
RIB Source code for this example
if( mod(s*frequency,1) < density ||
mod(t*frequency,1) < density )
Oi = 1.0;
else
Oi = 0.0;
Ci = Oi * ( Cs * ( Ka*ambient() + Kd*diffuse(Nf) ) +
specularcolor*Ks* specular(Nf,V,roughness));
}
The local variable V is defined to be the normalized vector for the
incident light rays direction. The incident light ray direction is the
direction from which the camera views the current surface coordinate.
This value is used later to compute the specular highlight to be used on
the portion of the surface which will not be cut out of the sphere.
The next thing the shader does is to compute the modulo of the s component
of the texture space times the frequency of the grid lines of the
wireframe. This value is always less than 1 (the modulo of s*frequency is
the remainder left for n*1 < s*frequency for some value n). If this value
is also less then the density then the current coordinate on the surface is
part of the visible wireframe that traverses the surface horizontally.
Likewise, the same modulo is computed for t*frequency and if this value is
also less than the density then the current coordinate point is on one of
the visible verticle grid lines of the wireframe. Any point for which the
module of either of these is greater than the density is rendered
completely transparent. The last line computes the grid lines based on the
current surface color and a slightly metallic lighting model.
The default value for the density is .25, which means that approximately
1/4 of the surface will be visible wireframe. Changing the value with an
instance variable to .1 would cause the the wireframe grid lines to become
thinner. Figure 7 shows an example of this. Changing the frequency to a
smaller number would cause fewer grid lines to be rendered.
While working on my entry for the March/April 1997 round of the IRTC I
wrote my first shader - a shader to simulate 3 holed notebook paper. This
simplistic shader offers some of the characteristics of the previous
examples in producing regularly spaced horizontal and verticle lines plus
the added feature of fully transparent circular regions that are positioned
by instance variables.
We start by defining the parameters needed by the shader. There are quite
a few more parameters than the other shaders. The reason for this is that
this shader works on features which are not quite so symmetrical. You can
also probably chalk it up to my inexperience.
color hcolor = color "rgb" (0, 0, 1);
color vcolor = color "rgb" (1, 0, 0);
float hfreq = 34;
float vfreq = 6;
float skip = 4;
float paper_height = 11;
float paper_width = 8.5;
float density = .03125;
float holeoffset = .09325;
float holeradius = .01975;
float hole1 = 2.6;
float hole2 = 18;
float hole3 = 31.25;
The colors of the horizontal and vertical lines come first. There are, by
default, 34 lines on the paper with the first 4 "skipped" to give the small
header space at the top of the paper. The vertical frequency is used to
divide the paper in n equal vertical blocks across the page. This is used
to determine the location of the single verticle stripe. We'll look at
this again in a moment.
The paper height and width are used to map the parameter space into the
correct dimensions for ordinary notebook paper. The density parameter is
the width of each of the visible lines (horizontal and vertical) on the
paper. The hole offset defines the distance from the left edge of the
paper to the center point of the 3 holes to be punched out. The holeradius is
the radius of the holes and the hole1-hole3 parameters give the horizontal
line over which the center of that hole will live. For example, for hole1
the center of the hole is 2.6 horizontal stripes down. Actually, the
horizontal stripes are created at the top of equally sized horizontal
blocks, and the hole1-hole3 values are number of horizontal blocks to
traverse down the paper for the holes center.
Now lets look at how the lines are created.
The smoothstep() function is part of the standard RenderMan functions
and returns a value that is between 0 and 1, inclusive, that shows where
"tt" sits between the min and max values. If this value is not at
either end then the current surface point lies in the bounds of a
horizontal line. The point is given the "hcolor" value mixed with the
current surface color We mix the colors in order to allow
the edges of the lines to flow smoothly between the horizontal lines
color and the color of the paper. In other words, this allows for
antialiasing the horizontal lines. The problem with this is - it doesn't
work. It only aliases one side of the line, I think. In any case, you can
see from Figure 8 that the result does not quite give a smooth,
solid set of lines.
surface_color = Cs;
This line simply initializes a local variable to the current color of
the surface. We'll use this value in computing a new surface color
based on whether the point is on a horizontal or vertical line.
/*
* Layer 1 - horizontal stripes.
* There is one stripe for every
* horizontal block. The stripe is
* "density" thick and starts at the top of
* each block, except for the first "skip"
* blocks.
*/
tt = t*paper_height;
for ( horiz=skip; horiz<hfreq; horiz=horiz+1 )
{
min = horiz*hblock;
max = min+density;
val = smoothstep(min, max, tt);
if ( val != 0 && val != 1 )
surface_color = mix(hcolor, Cs, val);
}
This loop runs through all the horizontal blocks on the paper
(defined by the hfreq parameter) and determines if the point
lies between the top of the block and the top of the block plus
the width of a horizontal line (specified with the density parameter).
Figure 8
RIB Source code for this example
Figure 8
An alternative approach would be to change the mix() function call (which
is part of the RenderMan shading lanague standard functions) to a more
simple mixture of the line color with the value returned by smoothstep().
This code would look like this:
min = horiz*hblock;
max = min+density;
val = smoothstep(min, max, tt);
if ( val != 0 && val != 1 )
surface_color = val*hcolor;
Alternatively, the line color could be used on its own, without combining
it with the value returned from the smooth step. This gives a very jagged
line, but the line is much darker even when used with smaller line
densities. The result from using the line color alone (with a smaller line
density) can be seen in Figure 9.
/* Layer 2 - vertical stripe */
ss = s*paper_width;
min = vblock;
max = min+density;
val = smoothstep(min, max, ss);
if ( val != 0 && val != 1 )
surface_color = mix(vcolor, Cs, val);
This next bit of code does exactly the same as the previous code
except it operates on the vertical line. Since there is only one
verticle line there is no need to check every vertical block, only
the one which will contain the visible stripe (which is specified
with the vblock parameter).
Finally we look at the hole punches. The center of the holes are computed
relative to the left edge of the paper:
shole = holeoffset*paper_width;
ss = s*paper_height;
tt = t*paper_height;
pos = (ss,tt,0);
Note that we use the papers height for converting the ss,tt variables into
the scale of the paper width and height. Why? Because if we used the
width for ss we would end up with eliptical holes. There is probably a
better way to deal with this problem (of making the holes circular) but
this method worked for me.
For each hole, the current s,t coordinates distance from the hole
centers is computed. If the distance is less than the holes radius then
the opacity for the incident ray is set to completely transparent.
/* First Hole */
thole = hole1*hblock;
hpos = (shole, thole, 0);
Oi = filterstep (holeradius*paper_width,
distance(pos,hpos));
/* Second Hole */
thole = hole2*hblock;
hpos = (shole, thole, 0);
Oi *= filterstep (holeradius*paper_width,
distance(pos,hpos));
/* Third Hole */
thole = hole3*hblock;
hpos = (shole, thole, 0);
Oi *= filterstep (holeradius*paper_width,
distance(pos,hpos));
Filterstep is, again, a standard function in the RenderMan specification.
However, this function was not documented by either the RenderMan Interface
Specification or the RenderMan Companion. According to Larry Gritz
The filterstep() function is identical to step, except that it is
analytically antialiased. Similar to the texture() function,
filterstep actually takes the derivative of its second argument, and
"fades in" at a rate dependent on how fast that variable is changing.
In technical terms, it returns the convolution of the step function
with a filter whose width is about the size of a pixel. So, no
jaggies.
Thus, using filterstep() helped to antialias the edges of the holes
(although its not that obvious from such a small image given in Figures 8
and 9). I didn't try it, but I bet filterstep() could probably be used to
fix the problems with the horizontal and vertical lines.
This simple texture map example is used in my Post Detention image
which I entered in the March/April 1997 IRTC. The actual shader is taken
from the archive collection by Guido Quaroni, and the shader originally
comes from Larry Knott (who I presume works at Pixar). I didn't add an
image of this since all you would see would be the original image mapped on a
flat plane, which really doesn't show anything useful. If you want to take
a look at the chalkboard in a complete scene, take a look at the
companion article
in this months Graphics Muse column.
Like the other shader examples, this one is fairly straightforward. An
image filename is passed in the texturename parameter. Note that
image files must be TIFF files for use with BMRT. The texture coordinates
are used to grab a value from the image file which is then combined with
the ambient and diffuse lighting for the incident ray. If a specular
highlight has been specified (which it is by default in the Ks parameter)
then a specular highlight is added to the incident ray. Finally, the
output value, Ci, is combined with the surfaces opacity for the final color
to be used by the current surface point.
We've already seen an example of displacement maps using the threads()
shader. Lets take a quick look at the shader code:
magnitude = (sin( PI*2*(t*frequency +
s + phase))+offset) * Km;
Here, the displacement of the surface point is determined by using a
phased sinusoidal. The t variable determines the position lengthwise
across the surface and s is used to cause the spiraling effect. The next
bit of code
if( t > (1-dampzone))
magnitude *= (1.0-t) / dampzone;
else if( t < dampzone )
magnitude *= t / dampzone;
causes the ends of the surface, in our case a cylinder, to revert to the
original shape. For our example that means this forces the shader to leave
the ends circular. This helps to keep the object that has been threaded in
a shape that is easily joined to other objects. In the RenderMan
Companion, the threaded cylinder is joined to a glass bulb to form a
light bulb. Finally, the last two lines
P += normalize(N) * magnitude;
N = calculatenormal(P);
cause the point to be moved and the normal for the new point to be
calculated. In this way the point visually appears to have moved, which
indeed it has.
© 1996 by Michael J. Hammel
"Linux Gazette...making Linux just a little more fun!"
Kandinski
By Jeff Hohensee,
ott@casper.com
How Kandinski operates ( I hope )
Copy a MIDI file with some tonal music to filename in.mid .
Run your ANSI Forth in the same directory. Include the Kandinski code into
your dictionary. Type main at the ok prompt. Kandinski will check in.mid
for a MIDI header. If in.mid is a midi file, Kandinski will traverse
tracks until it finds a noteon message. It will then tell you a bit about
the track and ask you if you want to make a picture of it. Hit y and it
will ask you if you want to use a piano or an organ type volume envelope.
The piano option is curently just a stub. Kandinski will then ask you to
hit a key to seed the filename randomizer. Kandinski will then create a
picture file with a filename of the form kanrrrrr.ppm, where r is a random
letter. The track portion of the program repeats if there are more tracks
with notes. The pictures created by Kandinski are 640 by 80 pixels, 24
bits color depth. I will soon be putting some Kandinski output up at
http://cqi.com/~humbubba
( kandinski )
( ANSI Forth sourcecode Rick Hohensee begun 199703 )
( A MIDIfile-to-still-picture implementation of my Cycluphonic method
of correlating colors and musical pitches. )
( used i486 Slackware Linux from the InfoMagic LDR sept 96, pfe,
Jeff Glatt's MIDI docs, dpans7 )
( redistribution permission contingent on authorship credit )
( default number base of file is.... ) decimal
( app notes, pfe file-postition is a DOUBLE!
MIDI sizes are SINGLEs
YEESH! "f0" is a variable! AAAAARRRRGGG!!!
hex f0 decimal . doesn't work as wished. )
( my prefered tools, jigs and cheats )
: binary decimal 2 base ! ;
: .base base @ dup decimal . base ! ;
: walk ." " key drop ;
: 0s ( wipe data stack )
depth dup if 0 do drop loop else drop then ;
: paddump ( [ count --- ] counted dump from pad )
pad swap dump ;
( app related ....)
0 value deltasum
2variable trkend 0 0 trkend 2!
0 value dpp ( deltas per pixel )
create rgbs 640 3 * allot
0 value trk#
variable midifile
0 value pbmfile
create organstate 128 allot
organstate 128 0 fill ( pfe allot leaves an "allot" string in the alloted
space )
create 12state 12 allot
12state 12 0 fill
0 value redac
0 value greenac
0 value blueac
0 value backfoot
create cycle 0 , 7 , 2 , 9 , 4 , 11 , 6 , 1 , 8 , 3 , 10 , 5 ,
create wheelred 12 allot
255 c, 255 c, 255 c, 127 c, 0 c, 0 c, 0 c, 0 c, 0 c, 127 c, 255 c, 255 c,
create wheelgreen 12 allot
0 c, 127 c, 255 c, 255 c, 255 c, 255 c, 255 c, 127 c, 0 c, 0 c, 0 c, 0 c,
create wheelblue 12 allot
0 c, 0 c, 0 c, 0 c, 0 c, 127 c, 255 c, 255 c, 255 c, 255 c, 255 c, 127 c,
0 value fid
create ppm
ascii P c, ascii 6 c, 10 c, ascii 6 c, ascii 4 c, ascii 0 c,
bl c, ascii 8 c, ascii 0 c,
bl c, ascii 2 c, ascii 5 c, ascii 5 c,
: msboff 127 and ;
: openin ( opens a file called in.mid in current dir
which can then be referenced via midifile @ )
S" in.mid" r/w bin open-file drop midifile ! ;
: in.mid ( --- fid_of_in.mid ) ( poorly factored, ) midifile @ ;
: inpos ( --- 2inpos ) ( get file position in in.mid )
midifile @ file-position drop ( ior) ;
: inpeek ( [ count --- ] counted read from in.mid to pad )
pad swap
midifile @ read-file drop ;
: trksize ( --- trksize ) ( DOES move inpos )
( build a 32 bit track size cell from the WRONGendian value
, from body0 to body0 )
4 inpeek drop ( endianism translation )
pad c@ 24 lshift
pad 1 + c@ 16 lshift +
pad 2 + c@ 8 lshift +
pad 3 + c@ + ;
2variable prevpos
2variable starttrk 0 0 starttrk 2!
: filebound ( fid --- 0 if inside file )
dup >r file-position drop r> file-size drop 2swap d< ;
: hoptrk ( [ --- inbounds_flag ] body0 to next trk body0 )
trksize 8 + 0 inpos d+ in.mid reposition-file drop
in.mid filebound ;
0 value envelope
0 value noteons 0 value noteoffs
: hinybble 240 and ; ( f0 is a &$^%##%$ variable name! )
hex
0f constant lonybble
binary
: bit7 10000000 and ;
decimal
0 value delta
: bytein pad 1 in.mid read-file drop
1 <> if ( error) cr
." end of in.mid "
quit else pad c@ then ;
: bignum 0
begin bytein dup bit7
while
msboff swap 7 lshift +
repeat
swap 7 lshift + ;
: ignore ( n --- ) ( add n to inpos )
0 inpos d+ in.mid reposition-file drop ;
: ignoreto ( delimiter --- ) ( ignore filebytes to delimiter )
begin dup bytein = until drop ;
0 value moment
: mthd ( --- da position of MThD or fail )
77 ignoreto 84 ignoreto 104 ignoreto 100 ignoreto inpos ;
: mtrk 77 ignoreto 84 ignoreto 114 ignoreto 107 ignoreto inpos ;
: seed
." hit a key please " key
time&date 2drop drop + + + in.mid + ;
: 128to12 ( organstate to 12state, i.e. midinote#s to notename#s )
12state 12 0 fill
128 0 do
organstate i + c@ if
1 i 12 mod 12state + c!
then ( simple for now )
loop
;
: 12torgb 0 to redac 0 to greenac 0 to blueac
12 0 do
12state i + c@ if
i cells cycle + @
cells dup wheelred + @ redac + 2 / to redac
dup wheelgreen + @ greenac + 2 / to greenac
wheelblue + @ blueac + 2 / to blueac
then
loop ;
: orgtorgb ( pixel# --- )
128to12
12torgb
dup redac swap 3 * rgbs + c!
dup greenac swap 3 * 1 + rgbs + c!
blueac swap 3 * 2 + rgbs + c!
;
: reset ( --- ) ( actions on an FF status byte )
bytein case
0 of bignum ignore ." ff 00 ignored " endof
1 of ." text " bignum ignore endof
2 of ." copyright " bignum ignore endof
3 of ." trackname " bignum ignore endof
4 of ." inst name " bignum ignore endof
5 of ." lyric " bignum ignore endof
6 of ." flow marker " bignum ignore endof
7 of ." cue point, sample " bignum ignore endof
33 of 2 ignore ( port # ) endof
47 of ( ." last event of track " ) 1 ignore endof
81 of 4 ignore endof
84 of 6 ignore ." smte o/s ignored " endof
88 of 5 ignore ( time sig ) endof
( ." unknown reset ff thang " )
endcase ;
: sysex ( sysexbyte --- ) ( i.e. message with status hinyb of f )
dup case
240 of 247 ignoreto ." ignoring f0 to f7 " drop endof
241 of ." miditimecode, unsupported " drop endof
242 of ." song position pointer " drop endof
243 of ." song select " drop endof
244 of ." unimplemented f4 sysex " drop endof
245 of ." unimplemented f5 sysex " drop endof
246 of ." tune calibrate " drop endof
249 of ." unimplemented f9 sysex " drop endof
247 of ." discontinue f0/240 stream " drop endof
248 of ." midi clock " drop endof
250 of ." restart song " drop endof
251 of ." midi continue, flow " drop endof
252 of ." stop " drop endof
254 of ." active sense message " drop endof
253 of ." unimplemented fd sysex " drop endof
255 of reset endof
." impossible sysex "
endcase ;
: envelope? cr ." piano envelope or organ? (p=piano/other=organ) " key
ascii p = if -1 to envelope else 0 to envelope then ;
: message ( survey pass )
bytein dup hinybble case
128 of 2 ignore noteoffs 1 + to noteoffs drop endof
144 of noteons 1+ to noteons 2 ignore drop endof
160 of 2 ignore drop endof
176 of 2 ignore drop endof
192 of 2 ignore drop endof
208 of 2 ignore drop endof
224 of 2 ignore drop endof
240 of cr sysex endof
endcase ;
: pianooff ." pianooff " 2 ignore ;
: pianoon 2 ignore ;
: organoff 0 organstate bytein + c! 1 ignore ;
: organon -1 organstate bytein + c! 1 ignore ;
: messageagain ( processing pass )
bytein dup hinybble case
128 of envelope if pianooff else organoff then drop endof
144 of envelope if pianoon else organon then drop endof
160 of 2 ignore drop endof
176 of 2 ignore drop endof
192 of 2 ignore drop endof
208 of 2 ignore drop endof
224 of 2 ignore drop endof
240 of cr sysex endof
endcase ;
: random.kan ( create file[name] kan[random].ppm )
seed srand
ascii k pad c! ascii a pad 1 + c! ascii n pad 2 + c!
8 3 do 26 random 97 + i pad + c! loop
ascii . pad 8 + c! ascii p pad 9 + c! ascii p pad 10 + c!
ascii m pad 11 + c! ;
: makepic
random.kan
pad 12 r/w create-file drop to pbmfile ( new filename exists )
ppm 16 pbmfile write-file drop
80 0 do
rgbs 640 3 * pbmfile write-file drop
loop
;
: process
0 to deltasum 0 to noteons 0 to noteoffs
640 0 do ( i=pixel )
begin
( bignum backfoot )
bignum deltasum + to deltasum
messageagain
i dpp * deltasum >
while
repeat
( paint pixel )
i orgtorgb
loop
makepic
;
: survey ( a track )
inpos starttrk 2!
trksize 0 inpos d+ trkend 2!
0 to deltasum 0 to noteons 0 to noteoffs
begin
bignum deltasum + to deltasum
message
inpos trkend 2@ d<
while
repeat
;
: track survey
noteons if ." This track has notes.... "
cr ." noteons " noteons . ." noteoffs " noteoffs .
." MIDI clocks per pixel " deltasum 640 / dup to dpp .
cr ." wanna do a pic of this track? (y/other) " key ascii y = if
envelope?
starttrk 2@ in.mid reposition-file drop inpos d. walk
noteons . dpp if
process else ." less than one clock per pixel, no can do " walk then
then then
;
: typecheck
mthd
inpos 2dup 4 0 d= if ." apparent std MIDI seq file. Yay. "
else 16 0 d= if ." apparent RMID MIDI file. OK. " else
cr ." in.mid is apparently not a MIDI file " cr
." Copy MIDI file to be processed to in.mid " bye then then ;
: main 0 to trk#
openin typecheck
begin
trk# 1 + dup to trk#
mtrk
track
( bytein does a QUIT on end-of-file )
again
;
Copyright © 1997, Jeff Hohensee
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
1997 Linux Expo
By Jon "maddog" Hall
maddog@zk3.dec.com
Copyright © 1997, Jon "maddog" Hall
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
A Fresh Beginning: The Enlightenment Window Manager
By Larry Ayers,
layers@vax2.rainis.net
Introduction
Features and Appearance
Availability
Closing Thoughts
Copyright © 1997, Larry Ayers
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
Visual Music: The Linux Port of Cthugha
By Larry Ayers,
layers@vax2.rainis.net
Introduction
What Cthugha Does
Running Cthugha
Copyright © 1997, Larry Ayers
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
Updates and Corrections
By Larry Ayers,
layers@vax2.rainis.net
GV
Hello Larry!
I enjoyed reading your article, but there are two remarks I want to
make:
- Your screen capture is one of the one modified gv that works with
all Athena Widgets, including the standard one. These modifications
were made by me (although it wasn't very hard once I realized how
well Johannes separated the Xaw3d stuff from the rest).
It would have been better to have a screen capture using libXaw3d, as
that is the standard look and feel. The last statement about having
to have Xaw3d is not very convincing this way.
- There is a gv homepage now:
http://wwwthep.physik.uni-mainz.de/~plass/gv/
This page currently features gv version 3, which can no longer be
used without libXaw3d. The last version of gv supporting standard
Xaw was 2.9.4 which will soon be available on a debian archive site.
Version 3 is even better than version 2 with respect to look and feel
(one of the first really convincing applications using Xaw3d, IMO)
and an improved postscript scanner.
While I'm sure that it isn't possible to change/add to the article,
there could be a short notice in the next gazette.
Helmut
--
Helmut Geyer Helmut.Geyer@iwr.uni-heidelberg.de
public PGP key available : finger geyer@saturn.iwr.uni-heidelberg.de
FileRunner
# This is an example of user-defined commands. This file should be named
# cmds and placed in your ~/.fr directory. It will then be read by
# FileRunner at startup. Versions of FileRunner prior to 2.3 need to have
#the file named .fr_cmds and placed directly in the home directory.
# This list should contain all user-defined commands formatted as:
# { { <button-title> <procedure-name> } {..} {..} }
set config(usercommands) {
{ XEmacs xemacs }
}
#
proc xemacs { filelist srcdir destdir } {
cd $srcdir
# set l {}
foreach f $filelist {
exec gnuclient -q $f
}
}
(gnuserv-start)
in the file. What this button does is send
the files you've selected to an already-running XEmacs process (I usually have
one running in a different virtual desktop than the one FileRunner is using).
XEmacs will then open up a new frame in your current desktop with the file(s)
displayed in it. This is handy for browsing source code.
wm2 and wmx
Afterstep
Xvile
TkDesk
The Midnight Commander
XEmacs Update
Copyright © 1997, Larry Ayers
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
Slackware
By Sean Dreilinger,
sean@kensho.com
Contents:
Slackware Is Not For You (Or Maybe It Is)
A Quick History
Date: 7 Jan 1994 15:48:07 GMT
Date: 10 Jan 1994 04:57:41 GMT
Why, Then?
Slackware Pros and Cons
Slackware is old It's mature, widely available,
and the most widely installed
Linux distribution
Slackware lacks sexy administrative tools a la RedHat
You're free to add
other distributions such as
the RedHat package manager
Slackware includes bundled security holes
We know what some of the
vulnerabilities are and
volunteers have posted fixes
Donald Knuth complained about the fonts
Patrick Volkerding fixed the
fonts
Linus Torvalds uses another distribution Oh well
Slackware is assembled by Devil Worshippers
Satanist crackers (not SATAN
itself) will avoid your box
Slackware is no longer
This is a myth, Slackware is
developed actively maintained, sans
marketing hype
Slackware is not supported by a commercial vendor or sanctionaed user
group
Linux support is available
along with consultants, explained further
in the section on Commercial
Support
Slackware is not created by a committee or development team
Good. A system designed by one
accountable individual is
cohesive
http://sunsite.unc.edu/LDP/HOWTO/Distribution-HOWTO.html
Planning
Literacy Required
Hardware Compatibility
http://sunsite.unc.edu/LDP/HOWTO/Hardware-HOWTO.html
ftp://ftp.cdrom.com/pub/linux/slackware/patches/
Thinking Through Storage And File Systems
Multiple Operating Systems On One Hard Drive
ftp://lrcftp.epfl.ch/pub/linus/local/lilo/
An overview of LILO and how you can use it are easily gleaned
from the LILO Mini-HOWTO:
http://sunsite.unc.edu/LDP/HOWTO/mini/LILO/
Designing a File System To Use Multiple Partitions
http://www.pathname.com/fhs/
Designing a File System To Use Multiple Hard Drives
http://sunsite.unc.edu/mdw/HOWTO/mini/Partition/
Upgrade? Think Twice!
ftp://ftp.wsc.com/pub/freeware/linux/update.linux/
http://sunsite.unc.edu/LDP/
Select An Installation Method
CD-ROM
Party!
Boot Disks: Always a Good Thing
Prepare To Be Questioned (There Will Be a Quiz...)
Contingency Plan: Food For Thought
Slackware Setup Worksheet
1, 2, 3, 4, or 5
1, 2, 3, 4, or 5
1, 2, 3, or 4
1, 2, 3, 4, 5, 6, 7 8, 9, 10, 11, 12, or 13
IDE CD-ROM: Enter the device name that represents
your IDE CD-ROM drive. This will probably be one of
these (in the order of most to least likely):
/dev/hdb /dev/hdc /dev/hdd /dev/hde /dev/hdf
/dev/hdg /dev/hdh /dev/hda
Device name
2. /dev/scd1
slakware or slaktext
Any combination of a ap d e f k n q t tcl x xap xd xv y and other disk sets offered, separated by spaces
1, 2, 3, 4, 5, 6, or 7
0, 1, 2, or 3
Making Slackware Happen
Build Some Boot Disks
Slackware Boot Kernel Image Descriptions
aztech.i CD-ROM drives: Aztech CDA268-01A,
Orchid CD-3110, Okano/Wearnes,
CDD110, Conrad TXC, CyCDROM CR520,
CR540
bare.i (none, just IDE support)
cdu31a.i Sony CDU31/33a CD-ROM
cdu535.i Sony CDU531/535 CD-ROM
cm206.i Philips/LMS cm206 CD-ROM with cm260
adapter card
goldstar.i Goldstar R420 CD-ROM (sometimes sold
in a Reveal "Multimedia Kit")
mcd.i NON-IDE Mitsumi CD-ROM support
mcdx.i Improved NON-IDE Mitsumi CD-ROM
support
net.i Ethernet support
optics.i Optics Storage 8000 AT CD-ROM (the
"DOLPHIN" drive)
sanyo.i Sanyo CDR-H94A CD-ROM support
sbpcd.i Matsushita, Kotobuki, Panasonic,
CreativeLabs (Sound Blaster),
Longshine and Teac NON-IDE CD-ROM
support
xt.i MFM hard drive support
7000fast.s Western Digital 7000FASST SCSI support
Advansys.s AdvanSys SCSI support
Aha152x.s Adaptec 152x SCSI support
Adaptec 1542 SCSI support
Aha1740.s Adaptec 1740 SCSI support
Aha2x4x.s Adaptec AIC7xxx SCSI support (For these cards: AHA-274x, AHA-2842,
& AHA-2940, AHA-2940W, AHA-2940U, AHA-2940UW, AHA-2944D, AHA-2944WD,
& AHA-3940, AHA-3940W, AHA-3985, AHA-3985W)
Am53c974.s AMD AM53/79C974 SCSI support
Aztech.s All supported SCSI controllers, plus CD-ROM support for Aztech CDA268-01A,
Orchid CD-3110, Okano/Wearnes CDD110, Conrad TXC, CyCDROM CR520, CR540
Buslogic.s Buslogic MultiMaster SCSI support
Cdu31a.s All supported SCSI controllers, plus CD-ROM support for Sony CDU31/33a
Cdu535.s All supported SCSI controllers, plus CD-ROM support for Sony CDU531/535
Cm206.s All supported SCSI controllers, plus Philips/LMS cm206 CD-ROM with cm260 adapter card
Dtc3280.s DTC (Data Technology Corp) 3180/3280 SCSI support
Eata\_dma.s DPT EATA-DMA SCSI support (Boards such as PM2011, PM2021, PM2041,
& PM3021, PM2012B, PM2022, PM2122, PM2322, PM2042, PM3122, PM3222,
& PM3332, PM2024, PM2124, PM2044, PM2144, PM3224, PM3334.)
Eata\_isa.s DPT EATA-ISA/EISA SCSI support (Boards such as PM2011B/9X,
& PM2021A/9X, PM2012A, PM2012B, PM2022A/9X, PM2122A/9X, PM2322A/9X)
Eata\_pio.s DPT EATA-PIO SCSI support (PM2001 and PM2012A)
Fdomain.s Future Domain TMC-16x0 SCSI support
Goldstar.s All supported SCSI controllers, plus Goldstar R420 CD-ROM (sometimes sold
in a Reveal "Multimedia Kit")
In2000.s Always IN2000 SCSI support
Iomega.s IOMEGA PPA3 parallel port SCSI support (also supports the parallel
port version of the ZIP drive)
Mcd.s All supported SCSI controllers, plusstandard non-IDE Mitsumi CD-ROM support
Mcdx.s All supported SCSI controllers, plus enhanced non-IDE Mitsumi CD-ROM support
N53c406a.s NCR 53c406a SCSI support
N\_5380.s NCR 5380 and 53c400 SCSI support
N\_53c7xx.s NCR 53c7xx, 53c8xx SCSI support (Most NCR PCI SCSI controllers use this driver)
Optics.s All supported SCSI controllers, plus support for the Optics Storage 8000
AT CDROM (the "DOLPHIN" drive)
Pas16.s Pro Audio Spectrum/Studio 16 SCSI support
Qlog\_fas.s ISA/VLB/PCMCIA Qlogic FastSCSI! support (also supports the Control
Concepts SCSI cards based on the Qlogic FASXXX chip)
Qlog\_isp.s Supports all Qlogic PCI SCSI controllers, except the PCI-basic,
which the AMD SCSI driver supports
Sanyo.s All supported SCSI controllers, plus Sanyo CDR-H94A CD-ROM support
Sbpcd.s All supported SCSI controllers, plus Matsushita, Kotobuki, Panasonic,
CreativeLabs (Sound Blaster), Longshine and Teac NON-IDE CDROM support
Scsinet.s All supported SCSI controllers, plus full ethernet support
Seagate.s Seagate ST01/ST02, Future Domain TMC-885/950 SCSI support
Trantor.s Trantor T128/T128F/T228 SCSI support
Ultrastr.s UltraStor 14F, 24F, and 34F SCSI support
Ustor14f.s UltraStor 14F and 34F SCSI support
Boot Into Action
Slackware Setup Program
Welcome to Slackware Linux Setup
Is That All?
and wondering "What Next?"
Troubleshooting Difficult Deliveries
Slackware Installation FAQs
You Get What You Pay For (Commercial Support)
http://sunsite.unc.edu/LDP/HOWTO/Commercial-HOWTO.html
Basking In the Afterglow
Consider Reinstalling!
Install And Test Key Applications
Secure the System
durak login: root
root login refused on this terminal.
durak login:
Back Up
Copyright © 1997, Sean Dreilinger
Published in Issue 17 of the Linux Gazette, May 1997
"Linux Gazette...making Linux just a little more fun!"
Linux Installation Project
By Kendall G. Clark,
kclark@dal284.computek.net
Linux Installation Project
NTLUG Members
Organizations
Copyright © 1997, Kendall G. Clark
Published in Issue 17 of the Linux Gazette, May 1997
Copyright © 1997 Specialized Systems Consultants, Inc.
For information regarding copying and distribution of this material see the
Copying License.
Contents:
About This Month's Authors
Larry Ayers
Larry Ayers lives on a small farm
in northern Missouri, where he is currently engaged in building a
timber-frame house for his family. He operates a portable band-saw mill,
does general woodworking, plays the fiddle and searches for rare
prairie plants, as well as growing shiitake mushrooms. He is also
struggling with configuring a Usenet news server for his local ISP.
Kendall G. Clark
Kendall Clark is a PHD candidate in systematic theology at
Southern Methodist University. He
is hard at work on his dissertation using Red Hat 4.1, LaTeX,
and AucTeX & Xemacs. He helped found NTLUG in the summer of 1996
with Stephen Denny and Tim Jones and currently serves as Acting
President. He makes his home with his wife Hope in Dallas,
Texas.
Jim Dennis
Jim Dennis
is the proprietor of
Starshine Technical Services.
His professional experience includes work in the technical
support, quality assurance, and information services (MIS)
departments of software companies like
Quarterdeck,
Symantec/
Peter Norton Group, and
McAfee Associates -- as well as
positions (field service rep) with smaller VAR's.
He's been using Linux since version 0.99p10 and is an active
participant on an ever-changing list of mailing lists and
newsgroups. He's just started collaborating on the 2nd Edition
for a book on Unix systems administration.
Jim is an avid science fiction fan -- and was
married at the World Science Fiction Convention in Anaheim.
Sean Dreilinger
Sean Dreilinger
suffered through two years of Los Angeles smog for a Masters degree in
library/information systems at UCLA. Linux swept him off his feet in grad
school and turned into a Network Administration career for the University.
Consulting on Internet strategy and info-system design in assorted
bored-rooms followed. Today he beams-in to www.interactivate.com from a
remote mountain cabin near Cuyamaca, California and is only required to
show his face at work once a week--nice job for the outdoors-loving and
socially inept. He lives with his lover Kathy and this incredible
high-altitude silence--punctuated only by the sound of wind rustling in the
Manzanita trees, hummingbirds fighting for a perch on the feeder, and that
reassuring whir of SCSI drives dancing with Linux under the desk. More life
story with explicit photos can be found at http://www.interactivate.com/people/sean/.
Jon "maddog" Hall
Jon "maddog" Hall is Senior Leader of Digital UNIX
Base Product Marketing, Digital Equipment Corporation.
Michael J. Hammel
Michael J. Hammel,
is a transient software engineer with a background in
everything from data communications to GUI development to Interactive Cable
systems--all based in Unix. His interests outside of computers
include 5K/10K races, skiing, Thai food and gardening. He suggests if you
have any serious interest in finding out more about him, you visit his home
pages at http://www.csn.net/~mjhammel. You'll find out more
there than you really wanted to know.
Rick Hohensee
Rick Hohensee is a guitar bum and former construction executive who has
so many irons in the fire he can't keep the fire going. Visit him on
the web at http://cqi.com/~humbubba.
Mike List
Mike List is a father of four teenagers, musician, printer (not
laserjet), and recently reformed technophobe, who has been into computers
since April,1996, and Linux since July.
Jesper Pedersen
Jesper Pedersen lives in Odense, Denmark, where he has studied computer science
at Odense University since 1990. He expects to obtain his degree in a year and
a half. He has a great job as a system manager at the
university, and also teaches computer science two hours a week. He is very
proud of his "child," The Dotfile Generator, which he wrote as part of his
job at the university. The idea for it came a year and a half ago, when he had
to learn how to configure Emacs by reading about 700 pages of the lisp manual.
It started small, but as time went by, it expanded into a huge project.
In his spare time, he does Yiu-Yitsu, listens to music, drinks beer and has fun with
his girl friend. He loves pets, and has a 200 litre aquarium and two very cute
rabbits.
Jay Painter
Jay Painter is the Systems Administrator at SSC.
Not Linux
Editor, Linux Gazette gazette@ssc.com
This page written and maintained by the Editor of Linux Gazette,
gazette@ssc.com