Updates from November, 2007 Toggle Comment Threads | Keyboard Shortcuts

  • Roberto Galoppini 5:58 pm on November 14, 2007 Permalink | Reply  

    Internet Governance Forum: transcripts from Rio 

    The Government of Brazil is hosting in Rio de Janeiro the second Internet Governance Forum meeting. The following is the output of the real-time captioning taken during the IGF.

    HELOISA MAGALHÃES: Good morning, ladies and gentlemen. I’m going to talk in Portuguese. Good morning, ladies and gentlemen. I am a Brazilian journalist from the “Valor Econômico,” and I am deeply honored to take part in this meeting. For us journalists in economics and finance, this issue is of utmost
    importance. And for me as a Brazilian, there’s special appeal to this. We are a country full of inequalities, and the Internet has proven to be a means of overcoming the challenge. First, I would like to call upon Mr. Ronaldo Lemos, who will chair the session. However, before, I’d like to remind all of you that our intention is to promote a debate. This is to be an interactive session.

    Questions and answers — questions from the audience. I would like to invite those who are sitting at the back of the room to come up closer so that we can have a true interactivity, so that we can have a more joyous interaction. First of all, Mr. Ronaldo, I give you the floor.

    Read the full transcript.

     
  • Roberto Galoppini 11:26 am on September 6, 2007 Permalink | Reply  

    FSF Europe: the beta version of the SELF Education Platform goes live 

    Yesterday SELF – a community-driven platform for the production and distribution of educational materials sponsored by the EU IST programme  – has been officially launched by the Free Software Foundation Europe in the Netherlands during a conference on Free Software in Education.

    The SELF platform aims to bring together educational institutes, training centres, software companies, publishing houses, government bodies and Free Software communities to centralise, create and disseminate educational and training materials on Free Software and Open Standards.

    From linuxelectrons:

    The SELF Platform has been developed by a global team of non-profit organisations, universities and volunteers engaged in the SELF Project, an initiative for the collaborative sharing and creation of free educational and training materials on Free Software and Open Standards. Users, primarily learners and teachers, are enabled to assemble selections of learning contents and create custom-made learning material for lessons in their language. The Platform is launched in beta stage to involve the growing community in optimising the tool.

    All SELF materials are available under the GNU Free Documentation License (GFDL), materials from third parties are licenses under various similar licenses.

    Let’s see now if  students and teachers will join the effort..

    Technorati Tags: Free Software Education, FSFE, IST, SELF

     
  • Carlo Daffara 1:51 pm on August 23, 2007 Permalink | Reply  

    Estimating the number of active and stable FLOSS projects 

    A recurring debate discussion among FLOSS-supporters and detractors is related to the estimation of the real number of active FLOSS projects. While it is easy to look at the main repository site (sourceforge.net) that boasts more than 100.000 projects, it is equally easy to look in more depth and realize that a significant number of those projects are really abandoned or have no significant development. How many active and stable projects are really out there?

    choicesToo many cereal choices by PartsNpieces

    For the purpose of obtaining some unbiased estimates in the context of the FLOSSMETRICS project, we performed a first search among the main repository sites and FLOSS announce portals; we also set a strict activity requirement, stately an activity index from 80 to 100% and at least a file release in the last 6 months. Of the overall 155959 projects, only 10656 (6.8%) are “active” (with a somehow very restrictive definition; a more relaxed release period of 1 year shows an active percentage of 9.2% or 14455 projects).

    However, while Sourceforge can rightly be considered the largest single repository, it is not the only potential source of projects; there are many other vertical repositories, among them BerliOS, Savannah, Gna! and many others, derived both from the original version of the Sourceforge code and many more based on a rewritten version called GForge. That gives a total of 23948 projects, to which (using a sampling of 100 projects from each) we have found a similar number of active projects (between 8% and 10%).

    The next step is the estimation of how many projects of the overall FLOSS landscape are hosted on those sites, and for performing this estimate we took the entire FreshMeat announce database, as processed by the FLOSSmole project and found that the projects that have an homepage in one of the repository sites are 23% of the total. This count is however biased by the fact that the probability of a project to be announced on FreshMeat is not equal for all projects; that is, english-based and oriented towards a large audience have a much higer probability to be listed. To take this into account, we performed a search for non-english based forges, and for software that is oriented towards a very specific area, using data from past IST projects like Spirit and AMOS.

    We have found that non-english projects are underrepresented in FreshMeat in a significant way, but as the overall “business-readiness” of those projects is unclear (as for example there may be no translations available, or be specific to a single country legal environment) we have ignored them. Vertical projects are also underrepresented, especially with regard to projects in scientific and technical areas, where the probability of being included is around 10 times lower compared to other kind of software. By using the results from Spirit, a sampling from project announcements in scientific mailing lists, and some repositories for the largest or more visible projects (like the CRAN archive, that hosts libraries and packages for the R language for statistics, that hosts 1195 projects) we have reached a lower bound estimate of around 12000 “vertical” and industry-specific projects. So, we have an overall lower bound estimate of around 195000 projects, of which we can estimate that 7% are active, leading to around 13000 active projects.

    Of those, we can estimate (using data from Slashdot, FreshMeat and the largest Gforge sites) that 36% fall in the “stable” or “mature” stage, leading to a total of around 5000 projects that can be considered suitable for an SME, that is with an active community, stable and with recent releases. It should be considered that this number is a lower bound, obtained with slightly severe assumptions; just enlarging the file release period from 6 months to one year nearly doubles the number of suitable projects. Also, this estimate does not try to assess the number of projects not listed in the announcement sites (even vertical application portals); this is a deliberate action, as it would be difficult to estimate the reliability of such a measure, and because the “findability” of a project and its probability of having a sustained community participation are lower if it is difficult to find information on the project in the first place; this means that the probability of such “out of the bounds” projects would probably be not a good opportunity for SME adoption in any case. By using a slightly more relaxed definition of “stability”, with an activity rating between 60% and 100% and at least a release in the last year, we obtain around 18000 stable and mature project from which to choose- not a bad result, after all.

    Technorati Tags: open source metrics, sourceforge, flossmetrics, flossmole

     
    • Bill Poser 6:57 am on August 27, 2007 Permalink

      The activity criterion used underestimates the number of projects that provide useful software. A project may not have had a recent release because it is complete and has no known bugs, or no bugs significant enough to fix. Of course, it would be difficult to take this into account without a lot more work since it would be necessary to examine the status of each project.

    • Carlo Daffara 7:47 am on August 28, 2007 Permalink

      As mentioned in the text, this is meant to provide a lower bound to the number of available, active and stable projects; as such, we have chosen a very strict definition of activity, and we used the project choice of “stability”, even considering that this lowers the number of suitable projects even more (there are many “beta” projects that are really stable). We already have found projects that are stable but not included in the count; an example is GNU make (that is stable, but having no new release in one year would not make it to the list).
      It must be considered, however, that even projects that are more or less finished (no more bugs) may need a small recompile or modification to adapt to changing platforms and environments; in this sense, stable project with no release in one year should be considered an exception and not the rule. Using a simple sampling approach, we estimate that those are less than 2% of our original count, and so we would not rise the package count in a significant way. Our main objective was to demonstrate that the lower bound of the number of both stable and maintained packages was significant, and I believe that that result was reached.
      Many thanks for your comment (and for reading the article thoroughly :-))

    • Nathan 9:24 pm on October 22, 2007 Permalink

      It would be very interesting to see that list of 18,000 stable, mature, active projects. Any plans to publish it?

    • Carlo Daffara 10:19 pm on October 29, 2007 Permalink

      For some of the forge sites that allows for data extraction, such a list can be obtained through the FLOSSMOLE data source. For those sites that have no search functionality, or that provide only part of their database in a searchable way, statistical methods were using based on a sampling approach, and in this case no list (just the numbers) can be obtained. It is important to understand that what we were looking at was a lower bound on the number of active and stable projects, not a “final” list.

    • Ali 4:43 pm on October 25, 2012 Permalink

      Hi,
      I am doing currently a research on open source firm. for statistical model i need number of projects registered to sourceforge year by year, is there any way to extract these information from sourceforge?

    • Roberto Galoppini 3:52 pm on October 27, 2012 Permalink

      Sure, look at the SourceForge Research Data at the Notre Dame University.

  • Carlo Daffara 9:54 am on August 6, 2007 Permalink | Reply  

    Open source collaboration: non-source code open projects 

    In the context of the joint research work with Roberto, I would like to present a small update in the OpenTTT project. OpenTTT is a EU-funded project (SSA-030595 INN7) that aims at bridging the separate worlds of technology transfer and open source software (OSS), by introducing novel methodologies for helping companies in the take up of technology.

    As part of the project, we are collecting examples of non-source code projects where collaboration or open licensing are critical, and prepared a listing of such activities. Such listing will be extended in the next weeks, also including previous work like the “Open Source Gift Guide” or a list of non software open source goods.

    As already discussed a large portion of work in OSS projects goes into non-code aspects, and as such should be investigated probably with the same interest that OSS commands today.

    Technorati Tags: openttt, EU projects

     
  • Roberto Galoppini 6:51 pm on July 6, 2007 Permalink | Reply  

    Internet Governance Forum: workshop proposals online for viewing&merging 

    The Internet Governance Workshop proposals submitted within the 30 June deadline have now been posted for viewing. Save the following dates:

    • Deadline for submitting proposal (abstracts + initial list of organizers ): 30 June.
      .
    • Completion of co-organizer and panellist arrangements and merge activities: July.
      .
    • Notification of selection – 31 July 2007.

    During July proponents of similar workshops will be encouraged to join forces and collaborate where it is feasible. Organizers of workshops are, therefore, expected to work with others who submit proposals on the same theme. A willingness to merge proposals is a requirement.
    The Government of Brazil will host in Rio de Janeiro on 12 – 15 November 2007 the second Internet Governance Forum meeting. The IGF website – run by the IGF Secretariat – supports the United Nations Secretary-General in carrying out the mandate from the World Summit on the Information Society with regard to convening a new forum for multi-stakeholder policy dialogue – namely the Internet Governance Forum.

    Technorati Tags: Internet Governance Forum, Open Consultation

     
  • Roberto Galoppini 5:46 pm on June 15, 2007 Permalink | Reply  

    Internet Governance Forum: updates for the next meeting 

    The Government of Brazil will host in Rio de Janeiro on 12 – 15 November 2007 the second Internet Governance Forum meeting. The IGF website – run by the IGF Secretariat – supports the United Nations Secretary-General in carrying out the mandate from the World Summit on the Information Society with regard to convening a new forum for multi-stakeholder policy dialogue – namely the Internet Governance Forum.
    Now a revised schedule for the Rio de Janeiro Meeting and a revised programme outline is available.

    Technorati Tags: Internet Governance Forum, Brasil, Open Consultation

     
  • Carlo Daffara 5:11 pm on May 21, 2007 Permalink | Reply  

    Open Source Adoption: OpenTTT, testing the IRC approach on open source 

    Choosing the best open source products is considered one of the biggest challenges in open source adoption. Software selection costs are so high that specialized consulting companies are doing it as their main job, see Optaros and Spikesource just to name two of. Why is it so difficult?

    Juggler Choose by Dovaneh

    There are many reasons:

    • there is no single place to search for OSS (sourceforge hosts a significant percentage of projects, but some merely started there and then moved elsewhere; there are many other forge-like sites and many software listing sites like freshmeat).
      .
    • there is no consistency in the software evaluation; even models like OSMM and BRR have many components that are based on human evaluation, and some more recent approaches even change the evaluation model and forms depending on the software area or market.
      .
    • there are many excellent projects that are not widely known; a great example is the large and sophisticated packages in the scientific software area, virtually unknown outside of a small community).

    This means that only a few projects get any visibility, and that many useful tools are not employed even when they could be the perfect match for a company. On this consideration, the EU funded a small project called OpenTTT, that tries to apply a “matching model” to help in the adoption process.

    It works like this:

    • A group of companies and public administrations are audited, and a set of needs in terms of software and IT functionalities are collected in structured forms (using a modification of the original IRC forms, called TR or technology requests);
      .
    • in parallel, OSS companies and developers are invited to fill a complementary form indicating on what projects they are offering services;
      .
    • requests are grouped, whenever possible, to find a single match for multiple companies;
      .
    • a manual matched process is performed to find potential matches between requests and offers matchmaking is perfected in one-to-one personal meetings at special “matchmaking” events;
      .
    • one has been recently performed at CeBIT and another at the CONFSL conference.

    An interesting twist of OpenTTT, that we hope to start soon, is the “club” concept. After all matches are performed, we expect that some needs will go unfulfilled; in this case we will try to find a “near match”, and try to group users with the same need into user clubs, and forward the information that an unfulfilled need has been identified to the groups of developers. After this, users and developers or companies are free to negotiate a commercial agreement, for example for implementing the missing pieces.

    See a chart depicting the process.

    I hope that this model can be a basis for a more structured and “grassroot” model for interaction between users and developers, not only because it gives an explicit recognition of the fact that OSS is not about price (at least not only about that) but also about flexibility and matching the user needs in a better way.

    Technorati Tags: OpenTTT, confsl, best practice, IRC

     
  • Carlo Daffara 6:40 am on May 18, 2007 Permalink | Reply  

    Open Source Blueprints: replicable experiments in open source adoption 

    Is there a better way for helping companies and public administrations in the OSS adoption process? Most adoptions are based on a few different paths, for example by grassroots adoption, from consultancy intervention, by trying to replicate a known success story. In this sense, the concept of “best practice” can be considered as a way to tell others of something that worked well, but in the past it has not been successful in replicating the experience.

    Best Practices Best Practices by andai

    So, considering that most public administrations are pushing for initiatives to help the adoption process (even if it mainly means creating another forge – like the Italian one just launched – I would like to propose the concept of the “implementation blueprint” as an
    extension of the best practice model. The idea came out of our experience in the
    Open TTT project, that is trying to leverage the technology transfer process used in the IRC network to facilitate the match between technology demand and offer in OSS.

    A blueprint is a replicable and complete description of a set of tools and processes that satisfied a specific need. In this sense, a complete blueprint must contain the following items:

    • a complete description of the needs; this should include a complete textual
      .
    • description of what was requested, including mandatory and secondary requests
      .
    • a description of the context of the needs, for example within a public
      .
    • administration, with specific legal requirements, an SME, etc
      .
    • the set of technologies used
      .
    • the process implemented
      .
    • criticalities or additional constraints appeared during the implementation process
      .
    • an estimate of the human effort invested in the migration process.

    Why so much detail? Because replicability requires a significant amount on information not only on the technological means, but also on how those tools were used to create a complete solution.

    As these mapping efforts are already under way – for example the Italian Open Source Observatory has a listing section, called “vetrina” that provides short summaries of public administrations’ experience with open source – it may be interesting to propose a collaborative writing process, maybe wikipedia-based, to turn nice-to-know stories into replicable experiences.

    [tags] Open Source Observatory, OpenTTT, best practice [tags]

     
  • Carlo Daffara 6:51 pm on May 15, 2007 Permalink | Reply  

    Trust networks, consultancies, and why proprietary market leaders are still leaders 

    Expanding the idea of peer conversations as basis for IT decisions, I would like to extend a little bit the reasons for my belief that this trend will probably continue, and lend to some unexpected results.

    Let’s start by thinking as a CIO that has to decide on a new technology, or in integrating a new software system in the company’s infrastructure. The only thing that the CIO knows is the fact that creating software from scratch is costly and requires a significant ongoing maintenance cost, so shifts the decision to a software platform from some vendor, and seeks advice on a company that may provide the necessary integration.

    Using this limited information, what the CIO knows is that:

    • there is a large number of potential platforms to choose from
      .
    • some may be more appropriate than others, and that choosing the wrong one may cause significant delays and added cost
      .
    • just browsing through the advertising material is not sufficient to choose in an appropriate way
      .
    • that the long-term viability of a company can only be guessed.

    So, what is the best strategy? We can try to imagine what a perfectly rational CIO would do, that is it would create a probability tree and try to guess at the potential events, their probability, and their impact. So, for example, if we choose by ourselves, the probabilities may be:

    graphic

    In this scenario, the CIO has to give an initial estimates at the probability of succeeding. How can she do it? By looking at similar tasks, for example. As most people uses Microsoft, or IBM, or SAP, she is fairly confident that she can use those too, and as those companies are still alive, they are probably doing it right. This is of course a false assumption, as there is limited information on failed or delayed project (outside the largest ones, like some government IT nightmares), but it is the only information that the CIO does have. Given this information, she knows that by choosing wisely, the potential cost of vendor A is 1.5, with vendor B is 2.4, with vendor C is 4. But she does not if the selection is appropriate, or if all the vendors have been included in the list.

    We have also not considered what happens after the end of the project, like what happens if the company leaves the market, or decides to change the platform without giving enough time for a migration strategy, and so on; but we will leave this for a later post.

    Now, let’s say that the CIO has already tried some projects, and discovered that she is unable to estimate probabilities with reasonable accuracy. At this point, she would probably go to a consultancy, that is an independent party with better information on the products, that has demonstrated to be able to select with more accuracy the appropriate probabilities. This is always advantageous, as long as the consultancy has an information advantage on the CIO; the price that she pays goes in a commensurate reduction in the risks associated with the project.

    But what happens when the consultancy seems no more able than the CIO to select the platform, or when it is suspected that the suggestions are not entirely independent? Then, the CIO has no alternative than trying as much as possible to remain on the “tried and tested”, and hope that everything will continue to be fine.

    What happened recently? The change is that the idea of openness and the availability of open forums allowed users to exchange information (sometimes even in an anonymous form), giving the CIO insight on what really works and what does not. This first hand information is for example what allowed many open source server projects to be deployed in a grass-roots fashion; because system administrators were exchanging information about them, and the best ones succeeded. Now this process is starting to be used at higher levels, and this goes back to the death of generalists conferences: as those do not allow for a venue for information exchange in a bilateral way, the users started feeling that it was not useful anymore when compared to web, second life, traditional marketing and so on.

    So we suppose that users (CIOs) are more interested in conversations. But can a CIO base her own opinion on talking with strangers? The reality is that in a way similar to how Google PageRanks adjusts relevance, the user networks created on blogs, digg-like social sites, or unconferences are adjusting themselves for relevance, and allows trust to emerge from seemingly untrusted parties.

    The concept is simple: let’s say that a user talks on a blog about his experience with a product, and other read about it. Around this post, may additional links may be created, some criticizing, some praising the text; and eventually, some users that share information often may become “daily reading material”. The usefulness, and reliability of the source can be inferred easily, by reading at the text itself, if the reasoning or the experience seems reasonable, and how others react to the post.

    While it may be imaginable that one blogger may be paid for talking in a positive way about a product, it is difficult to imagine that *every* user is biased or unreliable, and we can read and verify even the dissenting views with ease. This way, “reliable” writers and experts can emerge for free, and the CIO can verify everything without paying a consultant to get the same information. Of course this does not means that errors do not happen – only that errors are public, and that it is possible for everyone to check any step or any information against public sources.

    This is the real value that is arising from “web2.0” networks, that is the spontaneous creation of networks of peers, that can be trusted thanks to their transparency and willingness to cooperate. I can only guess that this form of value will be probably not be judged in a positive way by sellers, as this negates some lock-in advantage (the push for unified single-company platforms, for example); but this may be the only potential way to exit from a “lemon market” and giving back to the user the power to choose among products in an unbiased way.

    [trust networks, peer discovery, open source]

     
  • Carlo Daffara 7:03 pm on May 14, 2007 Permalink | Reply  

    Conferences, knowledge dissemination and the discovery of peers 

    As seen the traditional process used by companies to disseminate information and collect potential customers is becoming less and less useful; i is just the beginning of an overall transformation of how companies look at external information sources (like consulting companies).

    In the beginning of the commercial computer era, most users were connected through user clubs, since most software was developed in-house, and the software market was still in its infancy. Groups like SHARE, the first unix communities, VAX users groups and such provided the essential knowledge technicians needed, and were centered on the idea that software and hardware vendors were few, and user experiences were centered on real and concrete evidence.

    Unconference Unconference by MichaelBee

    With the consolidation of the shrinkwrapped software market and the multiplication of deployable technologies, the need for directions and information was not satisfiable with user conferences, and the consultancies were born – fundamentally, people with deep knowledge of a specific sector, reselling this knowledge to reduce the risk of implementation of a new technologies, or the time necessary to implement it. This period marked the beginning of comparison tools (like the infamous Quadrant), necessary in a world where one solution was exluding all the others.

    Open standards, open source and the substantial opening of IT architectures changed everything again; this, and the fact that consultancies were no longer current or reliable on trends that change in a very short time (anyone remembers the “push web” craze? the original Microsoft internet killer, Microsoft network? WAP?) and were found to be not so impartial after all.

    This void is being filled by a new generation of knowledge disseminator, be it small and efficient consultancies like RedMonk (that show that openness can be effective) and vertical conferences, that are less trade shows and more conversations. This resurgence of exchanging information as peers is what is really innovative, or maybe a return to the roots; the fact that customers are being treated less as passive suppliers of money, and more as partners in a long-term strategy, in a way that is strikingly similar to the kind of partnership that OSS companies create with their customers.

    Technorati Tags: Open Source conference, peers discovery, redmonk, knowledge dissemination

     
c
Compose new post
j
Next post/Next comment
k
Previous post/Previous comment
r
Reply
e
Edit
o
Show/Hide comments
t
Go to top
l
Go to login
h
Show/Hide help
shift + esc
Cancel