Alan McKinnon <alan.mckinnon <at> gmail.com> writes:
> I doubt dpkg and rpm aren't going to be much use to you, unless you > really want to run two package managers. Besides, both are not > especially useful with the front ends apt* and yum. I'd just use those to unpackage and maybe preprocess some of the codes. Agreed. I do not want a full blown deb or rpm package manager just a way to install and evaluate some of those codes before beginning a more arduous and comprehensive task. Maybe I should just put up a RH/centos box and evaluate codes there. It seems *everything* I want to test and look at in the cluster and hpc world, as a rpm or deb package; so I'm looking for a time saver, to surf thru the myriad of codes I'm getting; many look very cool from the outside, but once I run them, they are pigs....... Then a slick way to keep them secure and clean it out. Maybe I need chroot jails too? I spend way to much time managing codes rather than I do actually writing code. I feel confused often and cannot seem to master this git_thingy.... I have not code seriously in a long time and now it is becoming an obsession, but the old ways are draining my constitutional powers..... > Any special reason why you don't instead download the sources and build > them yourself with PREFIX=/usr/local ? Lots of errant codes flying everywhere so you have to pull a code audit to see what's in the raw tarballs before building. That takes way too much time. I'm working on setting up several more workstations for coding to isolate them from my main system. This approach you suggest is: error prone, takes too much time, and I'm lazy and sometimes even stupid. I need a structure methodology to be a one man extreme_hack_prolific system that prevents me from doing stupid things, whilst I'm distracted. Maybe I should just put up a VM resources on the net, blast tons of tests thru the vendors hardware and let them worry about the security ramifications? Some of it is these codes are based on 'functional languages' and I just do not trust what I do not fully understand. Stuff (files etc) goes everywhere and that makes me cautiously nervous. I have /usr/local for manual work and /usr/local/portage for ovelays (layman) but it's becoming a mess. There where to I put the work effort that is a result from repoman. Those codes seem to be parallel projects often when the code I'm evaluating needs to be cleaned up or extend to properly test. Furthermore I have a growing collection of file that result from kernel profiling via trace-cmd, valgrind, systemtap etc etc. As soon as I delete something, I need to re-generated it for one reason or another...... I just hope that this repo.conf effort helps be get more structurally organized? Did you see/test 'travis-ci' yet? [1] I'm not sure it's the same on github [2] but some of the devs are using it on github. James [1] http://docs.travis-ci.com/ [2] https://github.com/travis-ci/travis-ci